Min Pan, Teng Li, Chenghao Yang, Shuting Zhou, Shaoxiong Feng, Youbin Fang, Xingyu Li
{"title":"A Context-Aware BERT Retrieval Framework Utilizing Abstractive Summarization","authors":"Min Pan, Teng Li, Chenghao Yang, Shuting Zhou, Shaoxiong Feng, Youbin Fang, Xingyu Li","doi":"10.1109/WI-IAT55865.2022.00142","DOIUrl":null,"url":null,"abstract":"Recently, the multi-stage reranking framework based on pre-trained language model BERT can significantly improve the ranking performance on information retrieval tasks. However, most of these BERT-based reranking frameworks independently process query-chunk pairs and ignore cross-passages interaction. The context information around each candidate passage is extremely important for relevance judgement. Existing relevance aggregation methods obtain context information through statistical method and lost part of semantic information. Therefore, to capture this cross-passages interaction, this paper proposes a context-aware BERT ranking framework that utilizing abstractive summarization to enhance text semantics. By utilizing PEGASUS to summarize both sides of candidate passage accurately and then concatenate them as the input sequence, BERT could acquire more semantic information under the limitation of the input sequence’s length. The experimental results of two TREC data sets reveal the effectiveness of our proposed method in aggregating contextual semantic relevance.","PeriodicalId":345445,"journal":{"name":"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)","volume":"111 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WI-IAT55865.2022.00142","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Recently, the multi-stage reranking framework based on pre-trained language model BERT can significantly improve the ranking performance on information retrieval tasks. However, most of these BERT-based reranking frameworks independently process query-chunk pairs and ignore cross-passages interaction. The context information around each candidate passage is extremely important for relevance judgement. Existing relevance aggregation methods obtain context information through statistical method and lost part of semantic information. Therefore, to capture this cross-passages interaction, this paper proposes a context-aware BERT ranking framework that utilizing abstractive summarization to enhance text semantics. By utilizing PEGASUS to summarize both sides of candidate passage accurately and then concatenate them as the input sequence, BERT could acquire more semantic information under the limitation of the input sequence’s length. The experimental results of two TREC data sets reveal the effectiveness of our proposed method in aggregating contextual semantic relevance.