{"title":"基于前缀调优的机器阅读理解一致性学习框架","authors":"Yu Zhang , Bo Shen","doi":"10.1016/j.compeleceng.2025.110567","DOIUrl":null,"url":null,"abstract":"<div><div>Machine reading comprehension (MRC) is a traditional yet challenging natural language processing task. With advancements in pre-trained language models, several approaches have achieved remarkable results. However, the widespread existence of unanswerable questions presents a new challenge to the MRC task. In particular, language models trained on a universal corpus commonly suffer from hallucinations, and tend to fabricate plausible answers to unanswerable questions. Moreover, most existing methods perform task-specific fine-tuning based on pre-trained language models, necessitating the storage and updating of a substantial number of parameters. To address these issues, we propose a lightweight framework called Prefix-tuning-based Consistency Learning Machine Reading Comprehension (PCMRC) to mitigate model hallucination and improve the accuracy of answering unanswerable questions. PCMRC is structurally composed of two similar answer extractors and a prefix-tuning-based unanswerable discriminator. The answer extractors extract both a correct answer and a plausible answer, while the unanswerable discriminator generates a global probability indicating whether the question is unanswerable. A consistency training mechanism is then introduced to balance the probability of unanswerability with the two extracted answers. In this way, the proposed PCMRC mitigates the overconfidence of pre-trained language models, reduces the interference of unanswerable questions by plausible answers, and ultimately improves the accuracy of handling unanswerable questions. Experiments on the SQuAD2.0 dataset show that PCMRC achieves an F1-score of 88.9% and an accuracy of 89.2% for unanswerable questions. Notably, this method requires fewer than 1% of the parameters compared to fine-tuning, demonstrating its superiority and efficiency in the MRC task.</div></div>","PeriodicalId":50630,"journal":{"name":"Computers & Electrical Engineering","volume":"127 ","pages":"Article 110567"},"PeriodicalIF":4.9000,"publicationDate":"2025-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Prefix-tuning-based Consistency Learning Framework for Machine Reading Comprehension with unanswerable questions\",\"authors\":\"Yu Zhang , Bo Shen\",\"doi\":\"10.1016/j.compeleceng.2025.110567\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Machine reading comprehension (MRC) is a traditional yet challenging natural language processing task. With advancements in pre-trained language models, several approaches have achieved remarkable results. However, the widespread existence of unanswerable questions presents a new challenge to the MRC task. In particular, language models trained on a universal corpus commonly suffer from hallucinations, and tend to fabricate plausible answers to unanswerable questions. Moreover, most existing methods perform task-specific fine-tuning based on pre-trained language models, necessitating the storage and updating of a substantial number of parameters. To address these issues, we propose a lightweight framework called Prefix-tuning-based Consistency Learning Machine Reading Comprehension (PCMRC) to mitigate model hallucination and improve the accuracy of answering unanswerable questions. PCMRC is structurally composed of two similar answer extractors and a prefix-tuning-based unanswerable discriminator. The answer extractors extract both a correct answer and a plausible answer, while the unanswerable discriminator generates a global probability indicating whether the question is unanswerable. A consistency training mechanism is then introduced to balance the probability of unanswerability with the two extracted answers. In this way, the proposed PCMRC mitigates the overconfidence of pre-trained language models, reduces the interference of unanswerable questions by plausible answers, and ultimately improves the accuracy of handling unanswerable questions. Experiments on the SQuAD2.0 dataset show that PCMRC achieves an F1-score of 88.9% and an accuracy of 89.2% for unanswerable questions. Notably, this method requires fewer than 1% of the parameters compared to fine-tuning, demonstrating its superiority and efficiency in the MRC task.</div></div>\",\"PeriodicalId\":50630,\"journal\":{\"name\":\"Computers & Electrical Engineering\",\"volume\":\"127 \",\"pages\":\"Article 110567\"},\"PeriodicalIF\":4.9000,\"publicationDate\":\"2025-07-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computers & Electrical Engineering\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0045790625005105\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers & Electrical Engineering","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0045790625005105","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
Prefix-tuning-based Consistency Learning Framework for Machine Reading Comprehension with unanswerable questions
Machine reading comprehension (MRC) is a traditional yet challenging natural language processing task. With advancements in pre-trained language models, several approaches have achieved remarkable results. However, the widespread existence of unanswerable questions presents a new challenge to the MRC task. In particular, language models trained on a universal corpus commonly suffer from hallucinations, and tend to fabricate plausible answers to unanswerable questions. Moreover, most existing methods perform task-specific fine-tuning based on pre-trained language models, necessitating the storage and updating of a substantial number of parameters. To address these issues, we propose a lightweight framework called Prefix-tuning-based Consistency Learning Machine Reading Comprehension (PCMRC) to mitigate model hallucination and improve the accuracy of answering unanswerable questions. PCMRC is structurally composed of two similar answer extractors and a prefix-tuning-based unanswerable discriminator. The answer extractors extract both a correct answer and a plausible answer, while the unanswerable discriminator generates a global probability indicating whether the question is unanswerable. A consistency training mechanism is then introduced to balance the probability of unanswerability with the two extracted answers. In this way, the proposed PCMRC mitigates the overconfidence of pre-trained language models, reduces the interference of unanswerable questions by plausible answers, and ultimately improves the accuracy of handling unanswerable questions. Experiments on the SQuAD2.0 dataset show that PCMRC achieves an F1-score of 88.9% and an accuracy of 89.2% for unanswerable questions. Notably, this method requires fewer than 1% of the parameters compared to fine-tuning, demonstrating its superiority and efficiency in the MRC task.
期刊介绍:
The impact of computers has nowhere been more revolutionary than in electrical engineering. The design, analysis, and operation of electrical and electronic systems are now dominated by computers, a transformation that has been motivated by the natural ease of interface between computers and electrical systems, and the promise of spectacular improvements in speed and efficiency.
Published since 1973, Computers & Electrical Engineering provides rapid publication of topical research into the integration of computer technology and computational techniques with electrical and electronic systems. The journal publishes papers featuring novel implementations of computers and computational techniques in areas like signal and image processing, high-performance computing, parallel processing, and communications. Special attention will be paid to papers describing innovative architectures, algorithms, and software tools.