{"title":"CROSS-JEM:用于短文排序任务的准确高效交叉编码器","authors":"Bhawna Paliwal, Deepak Saini, Mudit Dhawan, Siddarth Asokan, Nagarajan Natarajan, Surbhi Aggarwal, Pankaj Malhotra, Jian Jiao, Manik Varma","doi":"arxiv-2409.09795","DOIUrl":null,"url":null,"abstract":"Ranking a set of items based on their relevance to a given query is a core\nproblem in search and recommendation. Transformer-based ranking models are the\nstate-of-the-art approaches for such tasks, but they score each query-item\nindependently, ignoring the joint context of other relevant items. This leads\nto sub-optimal ranking accuracy and high computational costs. In response, we\npropose Cross-encoders with Joint Efficient Modeling (CROSS-JEM), a novel\nranking approach that enables transformer-based models to jointly score\nmultiple items for a query, maximizing parameter utilization. CROSS-JEM\nleverages (a) redundancies and token overlaps to jointly score multiple items,\nthat are typically short-text phrases arising in search and recommendations,\nand (b) a novel training objective that models ranking probabilities. CROSS-JEM\nachieves state-of-the-art accuracy and over 4x lower ranking latency over\nstandard cross-encoders. Our contributions are threefold: (i) we highlight the\ngap between the ranking application's need for scoring thousands of items per\nquery and the limited capabilities of current cross-encoders; (ii) we introduce\nCROSS-JEM for joint efficient scoring of multiple items per query; and (iii) we\ndemonstrate state-of-the-art accuracy on standard public datasets and a\nproprietary dataset. CROSS-JEM opens up new directions for designing tailored\nearly-attention-based ranking models that incorporate strict production\nconstraints such as item multiplicity and latency.","PeriodicalId":501281,"journal":{"name":"arXiv - CS - Information Retrieval","volume":"13 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"CROSS-JEM: Accurate and Efficient Cross-encoders for Short-text Ranking Tasks\",\"authors\":\"Bhawna Paliwal, Deepak Saini, Mudit Dhawan, Siddarth Asokan, Nagarajan Natarajan, Surbhi Aggarwal, Pankaj Malhotra, Jian Jiao, Manik Varma\",\"doi\":\"arxiv-2409.09795\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Ranking a set of items based on their relevance to a given query is a core\\nproblem in search and recommendation. Transformer-based ranking models are the\\nstate-of-the-art approaches for such tasks, but they score each query-item\\nindependently, ignoring the joint context of other relevant items. This leads\\nto sub-optimal ranking accuracy and high computational costs. In response, we\\npropose Cross-encoders with Joint Efficient Modeling (CROSS-JEM), a novel\\nranking approach that enables transformer-based models to jointly score\\nmultiple items for a query, maximizing parameter utilization. CROSS-JEM\\nleverages (a) redundancies and token overlaps to jointly score multiple items,\\nthat are typically short-text phrases arising in search and recommendations,\\nand (b) a novel training objective that models ranking probabilities. CROSS-JEM\\nachieves state-of-the-art accuracy and over 4x lower ranking latency over\\nstandard cross-encoders. Our contributions are threefold: (i) we highlight the\\ngap between the ranking application's need for scoring thousands of items per\\nquery and the limited capabilities of current cross-encoders; (ii) we introduce\\nCROSS-JEM for joint efficient scoring of multiple items per query; and (iii) we\\ndemonstrate state-of-the-art accuracy on standard public datasets and a\\nproprietary dataset. CROSS-JEM opens up new directions for designing tailored\\nearly-attention-based ranking models that incorporate strict production\\nconstraints such as item multiplicity and latency.\",\"PeriodicalId\":501281,\"journal\":{\"name\":\"arXiv - CS - Information Retrieval\",\"volume\":\"13 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Information Retrieval\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.09795\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Information Retrieval","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09795","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
CROSS-JEM: Accurate and Efficient Cross-encoders for Short-text Ranking Tasks
Ranking a set of items based on their relevance to a given query is a core
problem in search and recommendation. Transformer-based ranking models are the
state-of-the-art approaches for such tasks, but they score each query-item
independently, ignoring the joint context of other relevant items. This leads
to sub-optimal ranking accuracy and high computational costs. In response, we
propose Cross-encoders with Joint Efficient Modeling (CROSS-JEM), a novel
ranking approach that enables transformer-based models to jointly score
multiple items for a query, maximizing parameter utilization. CROSS-JEM
leverages (a) redundancies and token overlaps to jointly score multiple items,
that are typically short-text phrases arising in search and recommendations,
and (b) a novel training objective that models ranking probabilities. CROSS-JEM
achieves state-of-the-art accuracy and over 4x lower ranking latency over
standard cross-encoders. Our contributions are threefold: (i) we highlight the
gap between the ranking application's need for scoring thousands of items per
query and the limited capabilities of current cross-encoders; (ii) we introduce
CROSS-JEM for joint efficient scoring of multiple items per query; and (iii) we
demonstrate state-of-the-art accuracy on standard public datasets and a
proprietary dataset. CROSS-JEM opens up new directions for designing tailored
early-attention-based ranking models that incorporate strict production
constraints such as item multiplicity and latency.