Ronghan Li , Dongdong Li , Haowen Yang , Xiaoxi Liu , Haoxiang Jin , RongCheng Pu , Qiguang Miao
{"title":"RECoT: Relation-enhanced Chains-of-Thoughts for knowledge-intensive multi-hop questions answering","authors":"Ronghan Li , Dongdong Li , Haowen Yang , Xiaoxi Liu , Haoxiang Jin , RongCheng Pu , Qiguang Miao","doi":"10.1016/j.neucom.2025.129903","DOIUrl":null,"url":null,"abstract":"<div><div>Open Domain question answering is designed to enable a computer to understand and answer any question on a wide range of topics. The prevalent retrieval-reading paradigm helps large language models (LLMs) when retrieving relevant text from external knowledge sources using questions, however the multi-hop question answering approach based on Chains-of-Thoughts (CoT) may perform poorly when it comes to complex questions. This is because there can be errors in generating sentences at each hop, and these errors accumulate, leading to significant deviations in the final result. In order to solve the above problems, this paper first extracted the relational triples of complex problems. Next, triples are used to select the most representative sentence at each step during CoT generation as the query for the next-hop retrieval.</div><div>The RECoT with GPT-3 results in significant improvements with F1 score up 5.1 points in downstream QA on 2WikiMultihopQA datasets and up 2.9 points on HotpotQA datasets. In addition, improvements in results can be obtained even with smaller models such as Flan-T5-large without additional training. In conclusion, RECoT reduced model hallucination and accelerated more accurate CoT reasoning to guide retrieval to get improved results. Code is publicly available at <span><span>https://github.com/XD-BDIV-NLP/RECoT</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"637 ","pages":"Article 129903"},"PeriodicalIF":5.5000,"publicationDate":"2025-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225005752","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Open Domain question answering is designed to enable a computer to understand and answer any question on a wide range of topics. The prevalent retrieval-reading paradigm helps large language models (LLMs) when retrieving relevant text from external knowledge sources using questions, however the multi-hop question answering approach based on Chains-of-Thoughts (CoT) may perform poorly when it comes to complex questions. This is because there can be errors in generating sentences at each hop, and these errors accumulate, leading to significant deviations in the final result. In order to solve the above problems, this paper first extracted the relational triples of complex problems. Next, triples are used to select the most representative sentence at each step during CoT generation as the query for the next-hop retrieval.
The RECoT with GPT-3 results in significant improvements with F1 score up 5.1 points in downstream QA on 2WikiMultihopQA datasets and up 2.9 points on HotpotQA datasets. In addition, improvements in results can be obtained even with smaller models such as Flan-T5-large without additional training. In conclusion, RECoT reduced model hallucination and accelerated more accurate CoT reasoning to guide retrieval to get improved results. Code is publicly available at https://github.com/XD-BDIV-NLP/RECoT.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.