Modeling rater judgments of interpreting quality: Ordinal logistic regression using neural-based evaluation metrics, acoustic fluency measures, and computational linguistic indices

Chao Han , Xiaolei Lu , Shirong Chen
{"title":"Modeling rater judgments of interpreting quality: Ordinal logistic regression using neural-based evaluation metrics, acoustic fluency measures, and computational linguistic indices","authors":"Chao Han ,&nbsp;Xiaolei Lu ,&nbsp;Shirong Chen","doi":"10.1016/j.rmal.2025.100194","DOIUrl":null,"url":null,"abstract":"<div><div>Human raters remain central to interpreting quality assessment (IQA); however, recent years have witnessed a growing body of research exploring automatic assessment. These studies have used machine translation evaluation metrics, acoustic fluency measures, and computational linguistic indices as separate approaches to model rater judgments of interpreting quality. Nonetheless, limited research has integrated these three types of measures within a single framework. To address this gap, this exploratory and proof-of-concept study adopts an integrative approach, combining these three types of measures to model rater judgments of interpreting quality as a classification problem. Using a dataset of 161 Chinese-to-English interpretations, we applied ordinal logistic regression analysis to identify significant predictors across fidelity, fluency, and linguistic dimensions. The analyses yielded two sets of significant predictors, including (a) COMET-22, mean length of unfilled pauses, mean length of run, and mean word length, and (b) BLEURT-20, phonation time ratio, speech rate, mean word length, type-token ratio for content words, type-token ratio for all words, and mean word frequency for content words. These models performed well on the testing dataset, particularly for classifying interpretations into four bands of overall interpreting quality (e.g., accuracy = .643, 1-off accuracy = .805), based on the Rasch-calibrated scores from human evaluation. These findings suggest that this integrated approach may enhance the precision and scalability of IQA and has the potential to reduce logistical burdens in large-scale professional interpreter certification exams and language proficiency tests.</div></div>","PeriodicalId":101075,"journal":{"name":"Research Methods in Applied Linguistics","volume":"4 1","pages":"Article 100194"},"PeriodicalIF":0.0000,"publicationDate":"2025-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Research Methods in Applied Linguistics","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2772766125000151","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Human raters remain central to interpreting quality assessment (IQA); however, recent years have witnessed a growing body of research exploring automatic assessment. These studies have used machine translation evaluation metrics, acoustic fluency measures, and computational linguistic indices as separate approaches to model rater judgments of interpreting quality. Nonetheless, limited research has integrated these three types of measures within a single framework. To address this gap, this exploratory and proof-of-concept study adopts an integrative approach, combining these three types of measures to model rater judgments of interpreting quality as a classification problem. Using a dataset of 161 Chinese-to-English interpretations, we applied ordinal logistic regression analysis to identify significant predictors across fidelity, fluency, and linguistic dimensions. The analyses yielded two sets of significant predictors, including (a) COMET-22, mean length of unfilled pauses, mean length of run, and mean word length, and (b) BLEURT-20, phonation time ratio, speech rate, mean word length, type-token ratio for content words, type-token ratio for all words, and mean word frequency for content words. These models performed well on the testing dataset, particularly for classifying interpretations into four bands of overall interpreting quality (e.g., accuracy = .643, 1-off accuracy = .805), based on the Rasch-calibrated scores from human evaluation. These findings suggest that this integrated approach may enhance the precision and scalability of IQA and has the potential to reduce logistical burdens in large-scale professional interpreter certification exams and language proficiency tests.
对口译质量的比较判断建模:使用基于神经的评价指标、声学流畅性测量和计算语言指标的有序逻辑回归
人类评分员仍然是口译质量评估(IQA)的核心;然而,近年来,越来越多的研究开始探索自动评估。这些研究使用机器翻译评估指标、声音流畅度测量和计算语言指标作为独立的方法来建立口译质量的模型。然而,有限的研究将这三种类型的措施整合在一个框架内。为了解决这一差距,本探索性和概念验证性研究采用了一种综合方法,将这三种类型的测量方法结合起来,对口译质量作为分类问题的评分判断进行建模。使用161个汉英口译数据集,我们应用有序逻辑回归分析来确定在保真度、流利度和语言维度上的显著预测因子。分析产生了两组重要的预测因子,包括(a) COMET-22,未填充停顿的平均长度,平均运行长度和平均单词长度,以及(b) BLEURT-20,发声时间比,语音率,平均单词长度,内容词的类型-标记比,所有单词的类型-标记比和内容词的平均词频。这些模型在测试数据集中表现良好,特别是将口译分为四个整体口译质量等级(例如,准确率= .643,1-off准确率= .805),基于人类评估的rasch校准分数。这些研究结果表明,这种综合方法可以提高IQA的准确性和可扩展性,并有可能减少大规模专业口译员认证考试和语言能力测试的后勤负担。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
4.10
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信