Method to assess the trustworthiness of machine coding at scale

IF 2.6 2区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH
Rebeckah K. Fussell, Emily M. Stump, N. G. Holmes
{"title":"Method to assess the trustworthiness of machine coding at scale","authors":"Rebeckah K. Fussell, Emily M. Stump, N. G. Holmes","doi":"10.1103/physrevphyseducres.20.010113","DOIUrl":null,"url":null,"abstract":"Physics education researchers are interested in using the tools of machine learning and natural language processing to make quantitative claims from natural language and text data, such as open-ended responses to survey questions. The aspiration is that this form of machine coding may be more efficient and consistent than human coding, allowing much larger and broader datasets to be analyzed than is practical with human coders. Existing work that uses these tools, however, does not investigate norms that allow for trustworthy quantitative claims without full reliance on cross-checking with human coding, which defeats the purpose of using these automated tools. Here we propose a four-part method for making such claims with supervised natural language processing: evaluating a trained model, calculating statistical uncertainty, calculating systematic uncertainty from the trained algorithm, and calculating systematic uncertainty from novel data sources. We provide evidence for this method using data from two distinct short response survey questions with two distinct coding schemes. We also provide a real-world example of using these practices to machine code a dataset unseen by human coders. We offer recommendations to guide physics education researchers who may use machine-coding methods in the future.","PeriodicalId":54296,"journal":{"name":"Physical Review Physics Education Research","volume":null,"pages":null},"PeriodicalIF":2.6000,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Physical Review Physics Education Research","FirstCategoryId":"95","ListUrlMain":"https://doi.org/10.1103/physrevphyseducres.20.010113","RegionNum":2,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"EDUCATION & EDUCATIONAL RESEARCH","Score":null,"Total":0}
引用次数: 0

Abstract

Physics education researchers are interested in using the tools of machine learning and natural language processing to make quantitative claims from natural language and text data, such as open-ended responses to survey questions. The aspiration is that this form of machine coding may be more efficient and consistent than human coding, allowing much larger and broader datasets to be analyzed than is practical with human coders. Existing work that uses these tools, however, does not investigate norms that allow for trustworthy quantitative claims without full reliance on cross-checking with human coding, which defeats the purpose of using these automated tools. Here we propose a four-part method for making such claims with supervised natural language processing: evaluating a trained model, calculating statistical uncertainty, calculating systematic uncertainty from the trained algorithm, and calculating systematic uncertainty from novel data sources. We provide evidence for this method using data from two distinct short response survey questions with two distinct coding schemes. We also provide a real-world example of using these practices to machine code a dataset unseen by human coders. We offer recommendations to guide physics education researchers who may use machine-coding methods in the future.

Abstract Image

大规模评估机器编码可信度的方法
物理教育研究人员对使用机器学习和自然语言处理工具从自然语言和文本数据(如对调查问题的开放式回答)中得出定量结论很感兴趣。他们的期望是,这种机器编码形式可能比人工编码更有效、更一致,从而可以分析更大、更广的数据集,而人工编码则不切实际。然而,使用这些工具的现有工作并没有研究无需完全依赖与人工编码进行交叉检查就能获得可信定量结果的规范,这有违使用这些自动化工具的初衷。在此,我们提出了一种由四个部分组成的方法,用于利用有监督的自然语言处理技术做出此类声明:评估训练有素的模型、计算统计不确定性、计算来自训练有素算法的系统不确定性,以及计算来自新数据源的系统不确定性。我们使用两种不同编码方案的两种不同简答调查问题的数据为这一方法提供了证据。我们还提供了一个使用这些方法对人类编码员未见过的数据集进行机器编码的实际例子。我们为将来可能使用机器编码方法的物理教育研究人员提供了指导建议。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Physical Review Physics Education Research
Physical Review Physics Education Research Social Sciences-Education
CiteScore
5.70
自引率
41.90%
发文量
84
审稿时长
32 weeks
期刊介绍: PRPER covers all educational levels, from elementary through graduate education. All topics in experimental and theoretical physics education research are accepted, including, but not limited to: Educational policy Instructional strategies, and materials development Research methodology Epistemology, attitudes, and beliefs Learning environment Scientific reasoning and problem solving Diversity and inclusion Learning theory Student participation Faculty and teacher professional development
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信