CROSS-JEM:用于短文排序任务的准确高效交叉编码器

Bhawna Paliwal, Deepak Saini, Mudit Dhawan, Siddarth Asokan, Nagarajan Natarajan, Surbhi Aggarwal, Pankaj Malhotra, Jian Jiao, Manik Varma
{"title":"CROSS-JEM:用于短文排序任务的准确高效交叉编码器","authors":"Bhawna Paliwal, Deepak Saini, Mudit Dhawan, Siddarth Asokan, Nagarajan Natarajan, Surbhi Aggarwal, Pankaj Malhotra, Jian Jiao, Manik Varma","doi":"arxiv-2409.09795","DOIUrl":null,"url":null,"abstract":"Ranking a set of items based on their relevance to a given query is a core\nproblem in search and recommendation. Transformer-based ranking models are the\nstate-of-the-art approaches for such tasks, but they score each query-item\nindependently, ignoring the joint context of other relevant items. This leads\nto sub-optimal ranking accuracy and high computational costs. In response, we\npropose Cross-encoders with Joint Efficient Modeling (CROSS-JEM), a novel\nranking approach that enables transformer-based models to jointly score\nmultiple items for a query, maximizing parameter utilization. CROSS-JEM\nleverages (a) redundancies and token overlaps to jointly score multiple items,\nthat are typically short-text phrases arising in search and recommendations,\nand (b) a novel training objective that models ranking probabilities. CROSS-JEM\nachieves state-of-the-art accuracy and over 4x lower ranking latency over\nstandard cross-encoders. Our contributions are threefold: (i) we highlight the\ngap between the ranking application's need for scoring thousands of items per\nquery and the limited capabilities of current cross-encoders; (ii) we introduce\nCROSS-JEM for joint efficient scoring of multiple items per query; and (iii) we\ndemonstrate state-of-the-art accuracy on standard public datasets and a\nproprietary dataset. CROSS-JEM opens up new directions for designing tailored\nearly-attention-based ranking models that incorporate strict production\nconstraints such as item multiplicity and latency.","PeriodicalId":501281,"journal":{"name":"arXiv - CS - Information Retrieval","volume":"13 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"CROSS-JEM: Accurate and Efficient Cross-encoders for Short-text Ranking Tasks\",\"authors\":\"Bhawna Paliwal, Deepak Saini, Mudit Dhawan, Siddarth Asokan, Nagarajan Natarajan, Surbhi Aggarwal, Pankaj Malhotra, Jian Jiao, Manik Varma\",\"doi\":\"arxiv-2409.09795\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Ranking a set of items based on their relevance to a given query is a core\\nproblem in search and recommendation. Transformer-based ranking models are the\\nstate-of-the-art approaches for such tasks, but they score each query-item\\nindependently, ignoring the joint context of other relevant items. This leads\\nto sub-optimal ranking accuracy and high computational costs. In response, we\\npropose Cross-encoders with Joint Efficient Modeling (CROSS-JEM), a novel\\nranking approach that enables transformer-based models to jointly score\\nmultiple items for a query, maximizing parameter utilization. CROSS-JEM\\nleverages (a) redundancies and token overlaps to jointly score multiple items,\\nthat are typically short-text phrases arising in search and recommendations,\\nand (b) a novel training objective that models ranking probabilities. CROSS-JEM\\nachieves state-of-the-art accuracy and over 4x lower ranking latency over\\nstandard cross-encoders. Our contributions are threefold: (i) we highlight the\\ngap between the ranking application's need for scoring thousands of items per\\nquery and the limited capabilities of current cross-encoders; (ii) we introduce\\nCROSS-JEM for joint efficient scoring of multiple items per query; and (iii) we\\ndemonstrate state-of-the-art accuracy on standard public datasets and a\\nproprietary dataset. CROSS-JEM opens up new directions for designing tailored\\nearly-attention-based ranking models that incorporate strict production\\nconstraints such as item multiplicity and latency.\",\"PeriodicalId\":501281,\"journal\":{\"name\":\"arXiv - CS - Information Retrieval\",\"volume\":\"13 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Information Retrieval\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.09795\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Information Retrieval","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09795","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

根据项目与给定查询的相关性对项目集进行排序是搜索和推荐中的一个核心问题。基于变换器的排名模型是此类任务的最先进方法,但它们对每个查询项的评分都是独立的,忽略了其他相关项的联合上下文。这导致了次优的排名准确性和高昂的计算成本。为此,我们提出了联合高效建模交叉编码器(Cross-encoders with Joint Efficient Modeling,CROSS-JEM),这是一种新颖的排序方法,它使基于变换器的模型能够对查询的多个项目进行联合评分,从而最大限度地提高参数利用率。CROSS-JEM 利用(a)冗余和标记重叠对多个项目(通常是搜索和推荐中出现的短文词组)进行联合评分,以及(b)对排名概率进行建模的新型训练目标。CROSS-JEM 达到了目前最先进的准确度,并且比标准交叉编码器的排序延迟时间低 4 倍以上。我们的贡献有三个方面:(i) 我们强调了排名应用对每次查询成千上万条项目的评分需求与当前交叉编码器有限能力之间的差距;(ii) 我们引入了 CROSS-JEM,用于对每次查询的多个项目进行联合高效评分;(iii) 我们在标准公共数据集和专有数据集上展示了最先进的准确性。CROSS-JEM 为设计基于早期注意力的定制排名模型开辟了新的方向,这些排名模型包含严格的生产约束条件,如项目多重性和延迟。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
CROSS-JEM: Accurate and Efficient Cross-encoders for Short-text Ranking Tasks
Ranking a set of items based on their relevance to a given query is a core problem in search and recommendation. Transformer-based ranking models are the state-of-the-art approaches for such tasks, but they score each query-item independently, ignoring the joint context of other relevant items. This leads to sub-optimal ranking accuracy and high computational costs. In response, we propose Cross-encoders with Joint Efficient Modeling (CROSS-JEM), a novel ranking approach that enables transformer-based models to jointly score multiple items for a query, maximizing parameter utilization. CROSS-JEM leverages (a) redundancies and token overlaps to jointly score multiple items, that are typically short-text phrases arising in search and recommendations, and (b) a novel training objective that models ranking probabilities. CROSS-JEM achieves state-of-the-art accuracy and over 4x lower ranking latency over standard cross-encoders. Our contributions are threefold: (i) we highlight the gap between the ranking application's need for scoring thousands of items per query and the limited capabilities of current cross-encoders; (ii) we introduce CROSS-JEM for joint efficient scoring of multiple items per query; and (iii) we demonstrate state-of-the-art accuracy on standard public datasets and a proprietary dataset. CROSS-JEM opens up new directions for designing tailored early-attention-based ranking models that incorporate strict production constraints such as item multiplicity and latency.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信