{"title":"英语写作中的错误检测与纠正:基于大语言模型的方法","authors":"Jin Wang","doi":"10.1016/j.aej.2025.08.005","DOIUrl":null,"url":null,"abstract":"<div><div>This study investigates the use of large language models (LLMs) for error detection and correction in English writing (EDCEW), focusing on the challenges of overcorrection and pedagogical feedback. We explore the relationship between targeted prompting strategies, transformer model architectures, and the error profiles of English language learners across CEFR proficiency levels (A, B, and C). We evaluated LLM through zero-shot, few-shot and fine-tuning approaches, showing that overcorrection is more common in texts from advanced learners, while lower proficiency levels show unique patterns of errors. To address this issue, we introduce a two-stage pipeline: first, a fine-tuned LLM leverages sequence-to-sequence modeling to perform precise token-level error detection and correction, achieving competitive F1 scores across multilingual datasets (English, German, Chinese); second, a prompted LLM (GPT 3.5) generates rule-based, natural language explanations for each identified error, enhancing the educational values. We present comprehensive benchmark results on standard WEDC datasets (BEA-2019, JFLEG), supported by human evaluations, that show the effectiveness of the proposed approach in producing accurate corrections and linguistically informative error explanations. The proposed model improves correction accuracy and facilitates deeper linguistic comprehension through explicit, pedagogically sound feedback, addressing a significant limitation of existing automated GEC tools. Moreover, this research advances LLM-driven WEDC by providing a robust framework that balances correction precision with pedagogical utility. Qualitative results further highlight the model’s ability to address diverse error types, such as Subject+Verb agreement, redundant comparisons, and incorrect preposition usage, with contextually appropriate corrections.</div></div>","PeriodicalId":7484,"journal":{"name":"alexandria engineering journal","volume":"129 ","pages":"Pages 1153-1164"},"PeriodicalIF":6.8000,"publicationDate":"2025-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"EDCEW-LLM: Error detection and correction in English writing: A large language model-based approach\",\"authors\":\"Jin Wang\",\"doi\":\"10.1016/j.aej.2025.08.005\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>This study investigates the use of large language models (LLMs) for error detection and correction in English writing (EDCEW), focusing on the challenges of overcorrection and pedagogical feedback. We explore the relationship between targeted prompting strategies, transformer model architectures, and the error profiles of English language learners across CEFR proficiency levels (A, B, and C). We evaluated LLM through zero-shot, few-shot and fine-tuning approaches, showing that overcorrection is more common in texts from advanced learners, while lower proficiency levels show unique patterns of errors. To address this issue, we introduce a two-stage pipeline: first, a fine-tuned LLM leverages sequence-to-sequence modeling to perform precise token-level error detection and correction, achieving competitive F1 scores across multilingual datasets (English, German, Chinese); second, a prompted LLM (GPT 3.5) generates rule-based, natural language explanations for each identified error, enhancing the educational values. We present comprehensive benchmark results on standard WEDC datasets (BEA-2019, JFLEG), supported by human evaluations, that show the effectiveness of the proposed approach in producing accurate corrections and linguistically informative error explanations. The proposed model improves correction accuracy and facilitates deeper linguistic comprehension through explicit, pedagogically sound feedback, addressing a significant limitation of existing automated GEC tools. Moreover, this research advances LLM-driven WEDC by providing a robust framework that balances correction precision with pedagogical utility. Qualitative results further highlight the model’s ability to address diverse error types, such as Subject+Verb agreement, redundant comparisons, and incorrect preposition usage, with contextually appropriate corrections.</div></div>\",\"PeriodicalId\":7484,\"journal\":{\"name\":\"alexandria engineering journal\",\"volume\":\"129 \",\"pages\":\"Pages 1153-1164\"},\"PeriodicalIF\":6.8000,\"publicationDate\":\"2025-08-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"alexandria engineering journal\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1110016825008750\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"alexandria engineering journal","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1110016825008750","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
EDCEW-LLM: Error detection and correction in English writing: A large language model-based approach
This study investigates the use of large language models (LLMs) for error detection and correction in English writing (EDCEW), focusing on the challenges of overcorrection and pedagogical feedback. We explore the relationship between targeted prompting strategies, transformer model architectures, and the error profiles of English language learners across CEFR proficiency levels (A, B, and C). We evaluated LLM through zero-shot, few-shot and fine-tuning approaches, showing that overcorrection is more common in texts from advanced learners, while lower proficiency levels show unique patterns of errors. To address this issue, we introduce a two-stage pipeline: first, a fine-tuned LLM leverages sequence-to-sequence modeling to perform precise token-level error detection and correction, achieving competitive F1 scores across multilingual datasets (English, German, Chinese); second, a prompted LLM (GPT 3.5) generates rule-based, natural language explanations for each identified error, enhancing the educational values. We present comprehensive benchmark results on standard WEDC datasets (BEA-2019, JFLEG), supported by human evaluations, that show the effectiveness of the proposed approach in producing accurate corrections and linguistically informative error explanations. The proposed model improves correction accuracy and facilitates deeper linguistic comprehension through explicit, pedagogically sound feedback, addressing a significant limitation of existing automated GEC tools. Moreover, this research advances LLM-driven WEDC by providing a robust framework that balances correction precision with pedagogical utility. Qualitative results further highlight the model’s ability to address diverse error types, such as Subject+Verb agreement, redundant comparisons, and incorrect preposition usage, with contextually appropriate corrections.
期刊介绍:
Alexandria Engineering Journal is an international journal devoted to publishing high quality papers in the field of engineering and applied science. Alexandria Engineering Journal is cited in the Engineering Information Services (EIS) and the Chemical Abstracts (CA). The papers published in Alexandria Engineering Journal are grouped into five sections, according to the following classification:
• Mechanical, Production, Marine and Textile Engineering
• Electrical Engineering, Computer Science and Nuclear Engineering
• Civil and Architecture Engineering
• Chemical Engineering and Applied Sciences
• Environmental Engineering