Grading Assistance for a Handwritten Thermodynamics Exam using Artificial Intelligence: An Exploratory Study

Gerd Kortemeyer, Julian Nöhl, Daria Onishchuk
{"title":"Grading Assistance for a Handwritten Thermodynamics Exam using Artificial Intelligence: An Exploratory Study","authors":"Gerd Kortemeyer, Julian Nöhl, Daria Onishchuk","doi":"arxiv-2406.17859","DOIUrl":null,"url":null,"abstract":"Using a high-stakes thermodynamics exam as sample (252~students, four\nmultipart problems), we investigate the viability of four workflows for\nAI-assisted grading of handwritten student solutions. We find that the greatest\nchallenge lies in converting handwritten answers into a machine-readable\nformat. The granularity of grading criteria also influences grading\nperformance: employing a fine-grained rubric for entire problems often leads to\nbookkeeping errors and grading failures, while grading problems in parts is\nmore reliable but tends to miss nuances. We also found that grading hand-drawn\ngraphics, such as process diagrams, is less reliable than mathematical\nderivations due to the difficulty in differentiating essential details from\nextraneous information. Although the system is precise in identifying exams\nthat meet passing criteria, exams with failing grades still require human\ngrading. We conclude with recommendations to overcome some of the encountered\nchallenges.","PeriodicalId":501565,"journal":{"name":"arXiv - PHYS - Physics Education","volume":"24 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - PHYS - Physics Education","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2406.17859","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Using a high-stakes thermodynamics exam as sample (252~students, four multipart problems), we investigate the viability of four workflows for AI-assisted grading of handwritten student solutions. We find that the greatest challenge lies in converting handwritten answers into a machine-readable format. The granularity of grading criteria also influences grading performance: employing a fine-grained rubric for entire problems often leads to bookkeeping errors and grading failures, while grading problems in parts is more reliable but tends to miss nuances. We also found that grading hand-drawn graphics, such as process diagrams, is less reliable than mathematical derivations due to the difficulty in differentiating essential details from extraneous information. Although the system is precise in identifying exams that meet passing criteria, exams with failing grades still require human grading. We conclude with recommendations to overcome some of the encountered challenges.
利用人工智能为手写热力学考试评分:一项探索性研究
我们以一次高风险的热力学考试(252 名学生,四个多部分问题)为样本,研究了人工智能辅助学生手写答案评分的四种工作流程的可行性。我们发现,最大的挑战在于将手写答案转换成机器可读的格式。评分标准的粒度也会影响评分效果:对整个问题采用细粒度的评分标准往往会导致记账错误和评分失败,而对部分问题进行评分则更可靠,但往往会遗漏细微差别。我们还发现,由于难以区分重要细节和无关信息,手绘图形(如流程图)的评分不如数学分析可靠。虽然该系统能准确识别符合合格标准的考试,但不及格的考试仍需要人工评分。最后,我们就如何克服遇到的一些挑战提出了建议。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信