Automated Answer Scoring for Engineering’s Open-Ended Questions

M. S. Ahmed
{"title":"Automated Answer Scoring for Engineering’s Open-Ended Questions","authors":"M. S. Ahmed","doi":"10.24297/ijrem.v10i0.8495","DOIUrl":null,"url":null,"abstract":"Audience Response System (ARS), like “clicker,” has proven their effectiveness in students’ engagement and in enhancing their learning. Apart from close-ended questions, ARS can help instructors to pose open-ended questions. Such questions are not scored automatically for that Automated Text Scoring; ATS is vastly used. This paper presents the findings of the development of an intelligent Automated Text Scoring, iATS, which provides instantaneous scoring of students’ responses to STEM-related factual questions. iATS is integrated with an Audience Response System (ARS), known as iRes, which captures students’ responses in traditional classrooms environment using smartphones. iATS Research is conducted to code and test three Natural Language Processing (NLP), text similarity methods. The codes were developed in PHP and Python environments. Experiments were performed to test Cosine similarity, Jaccard Index and Corpus-based and knowledge-based measures, (CKM), scores against instructor’s manual grades. The research suggested that the cosine similarity and Jaccard index are underestimating with an error of 22% and 26%, respectively. CKM has a low error (18%), but it is overestimating the score. It is concluded that codes need to be modified with a corpus developed within the knowledge domain and a new regression model should be created to improve the accuracy of automatic scoring.","PeriodicalId":247415,"journal":{"name":"International Journal of Research in Education Methodology","volume":"84 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Research in Education Methodology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.24297/ijrem.v10i0.8495","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Audience Response System (ARS), like “clicker,” has proven their effectiveness in students’ engagement and in enhancing their learning. Apart from close-ended questions, ARS can help instructors to pose open-ended questions. Such questions are not scored automatically for that Automated Text Scoring; ATS is vastly used. This paper presents the findings of the development of an intelligent Automated Text Scoring, iATS, which provides instantaneous scoring of students’ responses to STEM-related factual questions. iATS is integrated with an Audience Response System (ARS), known as iRes, which captures students’ responses in traditional classrooms environment using smartphones. iATS Research is conducted to code and test three Natural Language Processing (NLP), text similarity methods. The codes were developed in PHP and Python environments. Experiments were performed to test Cosine similarity, Jaccard Index and Corpus-based and knowledge-based measures, (CKM), scores against instructor’s manual grades. The research suggested that the cosine similarity and Jaccard index are underestimating with an error of 22% and 26%, respectively. CKM has a low error (18%), but it is overestimating the score. It is concluded that codes need to be modified with a corpus developed within the knowledge domain and a new regression model should be created to improve the accuracy of automatic scoring.
工程类开放式问题的自动答案评分
观众反应系统(ARS),就像“点击器”,已经证明了他们在学生参与和提高他们的学习方面的有效性。除了封闭式问题外,ARS还可以帮助教师提出开放式问题。这类问题不会在自动文本评分中自动得分;ATS被广泛使用。本文介绍了智能自动文本评分系统(iATS)的开发结果,该系统可以对学生对stem相关事实问题的回答进行即时评分。iATS集成了观众反应系统(ARS),即iRes,该系统使用智能手机捕捉学生在传统课堂环境中的反应。iATS研究对三种自然语言处理(NLP)文本相似度方法进行了编码和测试。这些代码是在PHP和Python环境中开发的。进行了余弦相似度、Jaccard指数、基于语料库和基于知识的测量(CKM)分数与教师手册分数的对比实验。研究表明,余弦相似度和Jaccard指数被低估,误差分别为22%和26%。CKM的误差较低(18%),但它高估了分数。结果表明,为了提高自动评分的准确性,需要在知识领域内建立相应的语料库对代码进行修改,并建立新的回归模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信