Evaluating the linguistic complexity of machine translation and LLMs for EFL/ESL applications: An entropy weight method

Yingqi Huang, Dechao Li, Andrew K.F. Cheung
{"title":"Evaluating the linguistic complexity of machine translation and LLMs for EFL/ESL applications: An entropy weight method","authors":"Yingqi Huang,&nbsp;Dechao Li,&nbsp;Andrew K.F. Cheung","doi":"10.1016/j.rmal.2025.100229","DOIUrl":null,"url":null,"abstract":"<div><div>English as a Foreign and Second Language (EFL/ESL) learners are increasingly using machine translation (MT) tools such as neural machine translations (NMTs) and large language models (LLMs) to enhance their language learning and translation processes due to their accuracy and efficiency in both cost and time compared with human translation. Given the distinct linguistic features exhibited by NMTs and LLMs, it is crucial to assess the linguistic complexity of texts produced by these tools to optimize their use in EFL/ESL teaching and learning. This study examines two forms of absolute linguistic complexity, namely lexical complexity and syntactic complexity, that influence EFL/ESL activities. Lexical complexity affects vocabulary recognition and semantic processing, while syntactic complexity influences sentence parsing and the internalization of grammatical rules. As both dimensions are multi-faceted and involve numerous indices that may vary in different directions (e.g., high values in certain measures and lower in others), an entropy weight method (EWM) is employed to assign data-driven weights and derive a balanced holistic complexity score. This approach enables a systematic comparison of translation outputs from NMTs (Google Translate, DeepL) and LLMs (ChatGPT-4o, OpenAI-o1). The findings reveal that LLMs generally exhibit higher holistic linguistic complexity, whereas NMTs tend to produce simpler translations. Pedagogically, LLM-translated texts may serve as more effective input for advanced language learners in EFL/ESL contexts, while NMT outputs may be more suitable for those with less linguistic proficiency.</div></div>","PeriodicalId":101075,"journal":{"name":"Research Methods in Applied Linguistics","volume":"4 3","pages":"Article 100229"},"PeriodicalIF":0.0000,"publicationDate":"2025-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Research Methods in Applied Linguistics","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2772766125000503","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

English as a Foreign and Second Language (EFL/ESL) learners are increasingly using machine translation (MT) tools such as neural machine translations (NMTs) and large language models (LLMs) to enhance their language learning and translation processes due to their accuracy and efficiency in both cost and time compared with human translation. Given the distinct linguistic features exhibited by NMTs and LLMs, it is crucial to assess the linguistic complexity of texts produced by these tools to optimize their use in EFL/ESL teaching and learning. This study examines two forms of absolute linguistic complexity, namely lexical complexity and syntactic complexity, that influence EFL/ESL activities. Lexical complexity affects vocabulary recognition and semantic processing, while syntactic complexity influences sentence parsing and the internalization of grammatical rules. As both dimensions are multi-faceted and involve numerous indices that may vary in different directions (e.g., high values in certain measures and lower in others), an entropy weight method (EWM) is employed to assign data-driven weights and derive a balanced holistic complexity score. This approach enables a systematic comparison of translation outputs from NMTs (Google Translate, DeepL) and LLMs (ChatGPT-4o, OpenAI-o1). The findings reveal that LLMs generally exhibit higher holistic linguistic complexity, whereas NMTs tend to produce simpler translations. Pedagogically, LLM-translated texts may serve as more effective input for advanced language learners in EFL/ESL contexts, while NMT outputs may be more suitable for those with less linguistic proficiency.
评估机器翻译和llm在EFL/ESL应用中的语言复杂性:一种熵权法
英语作为外语和第二语言(EFL/ESL)学习者越来越多地使用机器翻译(MT)工具,如神经机器翻译(nmt)和大型语言模型(llm),以提高他们的语言学习和翻译过程,因为与人工翻译相比,它们在成本和时间上都更加准确和高效。鉴于nmt和llm所表现出的不同语言特征,评估这些工具产生的文本的语言复杂性以优化它们在EFL/ESL教学和学习中的使用是至关重要的。本研究考察了绝对语言复杂性的两种形式,即词汇复杂性和句法复杂性,它们对EFL/ESL活动的影响。词汇复杂性影响词汇识别和语义加工,句法复杂性影响句子解析和语法规则内化。由于这两个维度都是多方面的,并且涉及许多指标,这些指标可能在不同的方向上变化(例如,某些措施的值较高,而其他措施的值较低),因此采用熵权法(EWM)来分配数据驱动的权重,并得出平衡的整体复杂性评分。这种方法可以对nmt(谷歌Translate, DeepL)和llm (chatgpt - 40, openai - 01)的翻译输出进行系统比较。研究结果表明,法学硕士通常表现出更高的整体语言复杂性,而nmt倾向于产生更简单的翻译。在教学上,法学硕士翻译的文本可以作为高级语言学习者在EFL/ESL环境中更有效的输入,而NMT输出可能更适合那些语言水平较低的人。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
4.10
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信