Xin Guan , Jiuxin Cao , Biwei Cao , Qingqing Gao , Bo Liu
{"title":"零射击常识问答的多跳常识知识注入框架","authors":"Xin Guan , Jiuxin Cao , Biwei Cao , Qingqing Gao , Bo Liu","doi":"10.1016/j.eswa.2025.129806","DOIUrl":null,"url":null,"abstract":"<div><div>Zero-shot commonsense question answering (QA) task is to evaluate the general reasoning ability of the language model without training on the specific datasets. The existing zero-shot framework transforms triples within the commonsense knowledge graphs (KGs) into QA-format samples, serving as a pre-training data source to integrate commonsense knowledge into the language model. However, this approach still faces the following challenges: 1) The model trained from synthetic QA generated from triples lacks the multi-hop commonsense knowledge required for handling complex QA problems. 2) Ambiguity caused by confusing commonsense knowledge within synthetic QA, making it challenging for models to discern semantically similar entities. To address the above problem, we propose a novel <strong>M</strong>ulti-hop <strong>C</strong>ommonsense <strong>K</strong>nowledge <strong>I</strong>njection Framework (MCKI). Specifically, we draw inspiration from human complex reasoning thinking and further propose a synthetic multi-hop commonsense QA generation method. Meanwhile, we introduce negative samples with high confusion in synthetic QA, and then use contrastive learning to improve the model’s ability to distinguish similar commonsense knowledge. Extensive experiments on five commonsense question answering benchmarks demonstrate that our framework achieves state-of-the-art performance, surpassing existing methods, including large language models like GPT3.5 and ChatGPT.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"298 ","pages":"Article 129806"},"PeriodicalIF":7.5000,"publicationDate":"2025-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-hop commonsense knowledge injection framework for zero-shot commonsense question answering\",\"authors\":\"Xin Guan , Jiuxin Cao , Biwei Cao , Qingqing Gao , Bo Liu\",\"doi\":\"10.1016/j.eswa.2025.129806\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Zero-shot commonsense question answering (QA) task is to evaluate the general reasoning ability of the language model without training on the specific datasets. The existing zero-shot framework transforms triples within the commonsense knowledge graphs (KGs) into QA-format samples, serving as a pre-training data source to integrate commonsense knowledge into the language model. However, this approach still faces the following challenges: 1) The model trained from synthetic QA generated from triples lacks the multi-hop commonsense knowledge required for handling complex QA problems. 2) Ambiguity caused by confusing commonsense knowledge within synthetic QA, making it challenging for models to discern semantically similar entities. To address the above problem, we propose a novel <strong>M</strong>ulti-hop <strong>C</strong>ommonsense <strong>K</strong>nowledge <strong>I</strong>njection Framework (MCKI). Specifically, we draw inspiration from human complex reasoning thinking and further propose a synthetic multi-hop commonsense QA generation method. Meanwhile, we introduce negative samples with high confusion in synthetic QA, and then use contrastive learning to improve the model’s ability to distinguish similar commonsense knowledge. Extensive experiments on five commonsense question answering benchmarks demonstrate that our framework achieves state-of-the-art performance, surpassing existing methods, including large language models like GPT3.5 and ChatGPT.</div></div>\",\"PeriodicalId\":50461,\"journal\":{\"name\":\"Expert Systems with Applications\",\"volume\":\"298 \",\"pages\":\"Article 129806\"},\"PeriodicalIF\":7.5000,\"publicationDate\":\"2025-09-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Expert Systems with Applications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0957417425034219\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417425034219","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Multi-hop commonsense knowledge injection framework for zero-shot commonsense question answering
Zero-shot commonsense question answering (QA) task is to evaluate the general reasoning ability of the language model without training on the specific datasets. The existing zero-shot framework transforms triples within the commonsense knowledge graphs (KGs) into QA-format samples, serving as a pre-training data source to integrate commonsense knowledge into the language model. However, this approach still faces the following challenges: 1) The model trained from synthetic QA generated from triples lacks the multi-hop commonsense knowledge required for handling complex QA problems. 2) Ambiguity caused by confusing commonsense knowledge within synthetic QA, making it challenging for models to discern semantically similar entities. To address the above problem, we propose a novel Multi-hop Commonsense Knowledge Injection Framework (MCKI). Specifically, we draw inspiration from human complex reasoning thinking and further propose a synthetic multi-hop commonsense QA generation method. Meanwhile, we introduce negative samples with high confusion in synthetic QA, and then use contrastive learning to improve the model’s ability to distinguish similar commonsense knowledge. Extensive experiments on five commonsense question answering benchmarks demonstrate that our framework achieves state-of-the-art performance, surpassing existing methods, including large language models like GPT3.5 and ChatGPT.
期刊介绍:
Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.