Towards adaptive support for self-regulated learning of causal relations: Evaluating four Dutch word vector models

IF 6.7 1区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH
Héctor J. Pijeira-Díaz, Sophia Braumann, Janneke van de Pol, Tamara van Gog, Anique B. H. de Bruin
{"title":"Towards adaptive support for self-regulated learning of causal relations: Evaluating four Dutch word vector models","authors":"Héctor J. Pijeira-Díaz,&nbsp;Sophia Braumann,&nbsp;Janneke van de Pol,&nbsp;Tamara van Gog,&nbsp;Anique B. H. de Bruin","doi":"10.1111/bjet.13431","DOIUrl":null,"url":null,"abstract":"<div>\n \n <section>\n \n <p>Advances in computational language models increasingly enable adaptive support for self-regulated learning (SRL) in digital learning environments (DLEs; eg, via automated feedback). However, the accuracy of those models is a common concern for educational stakeholders (eg, policymakers, researchers, teachers and learners themselves). We compared the accuracy of four Dutch language models (ie, spaCy medium, spaCy large, FastText and ConceptNet NumberBatch) in the context of secondary school students' learning of causal relations from expository texts, scaffolded by causal diagram completion. Since machine learning relies on human-labelled data for the best results, we used a dataset with 10,193 students' causal diagram answers, compiled over a decade of research using a diagram completion intervention to enhance students' monitoring of their text comprehension. The language models were used in combination with four popular machine learning classifiers (ie, logistic regression, random forests, support vector machine and neural networks) to evaluate their performance on automatically scoring students' causal diagrams in terms of the correctness of events and their sequence (ie, the causal structure). Five performance metrics were studied, namely accuracy, precision, recall, F<sub>1</sub> and the area under the curve of the receiver operating characteristic (ROC-AUC). The spaCy medium model combined with the neural network classifier achieved the best performance for the correctness of causal events in four of the five metrics, while the ConceptNet NumberBatch model worked best for the correctness of the causal sequence. These evaluation results provide a criterion for model adoption to adaptively support SRL of causal relations in DLEs.</p>\n </section>\n \n <section>\n \n <div>\n \n <div>\n \n <h3>Practitioner notes</h3>\n <p>What is already known about this topic\n\n </p><ul>\n \n <li>Accurate monitoring is a prerequisite for effective self-regulation.</li>\n \n <li>Students struggle to accurately monitor their comprehension of causal relations in texts.</li>\n \n <li>Completing causal diagrams improves students' monitoring accuracy, but there is room for further improvement.</li>\n \n <li>Automatic scoring could be used to provide adaptive support during diagramming.</li>\n </ul>\n <p>What this paper adds\n\n </p><ul>\n \n <li>Comparison of four Dutch word vector models combined with four machine learning classifiers for the automatic scoring of students' causal diagrams.</li>\n \n <li>Five performance metrics to evaluate the above solutions.</li>\n \n <li>Evaluation of the word vector models for estimating the semantic similarity between student and model answers.</li>\n </ul>\n <p>Implications for practice and/or policy\n\n </p><ul>\n \n <li>High-quality word vector models could (em)power adaptive support during causal diagramming via automatic scoring.</li>\n \n <li>The evaluated solutions can be embedded in digital learning environments (DLEs).</li>\n \n <li>Criteria for model adoption to adaptively support SRL of causal relations in DLEs.</li>\n \n <li>The increased saliency of (in)correct answers via automatic scoring might help to improve students' monitoring accuracy.</li>\n </ul>\n </div>\n </div>\n </section>\n </div>","PeriodicalId":48315,"journal":{"name":"British Journal of Educational Technology","volume":"55 4","pages":"1354-1375"},"PeriodicalIF":6.7000,"publicationDate":"2024-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bjet.13431","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"British Journal of Educational Technology","FirstCategoryId":"95","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/bjet.13431","RegionNum":1,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"EDUCATION & EDUCATIONAL RESEARCH","Score":null,"Total":0}
引用次数: 0

Abstract

Advances in computational language models increasingly enable adaptive support for self-regulated learning (SRL) in digital learning environments (DLEs; eg, via automated feedback). However, the accuracy of those models is a common concern for educational stakeholders (eg, policymakers, researchers, teachers and learners themselves). We compared the accuracy of four Dutch language models (ie, spaCy medium, spaCy large, FastText and ConceptNet NumberBatch) in the context of secondary school students' learning of causal relations from expository texts, scaffolded by causal diagram completion. Since machine learning relies on human-labelled data for the best results, we used a dataset with 10,193 students' causal diagram answers, compiled over a decade of research using a diagram completion intervention to enhance students' monitoring of their text comprehension. The language models were used in combination with four popular machine learning classifiers (ie, logistic regression, random forests, support vector machine and neural networks) to evaluate their performance on automatically scoring students' causal diagrams in terms of the correctness of events and their sequence (ie, the causal structure). Five performance metrics were studied, namely accuracy, precision, recall, F1 and the area under the curve of the receiver operating characteristic (ROC-AUC). The spaCy medium model combined with the neural network classifier achieved the best performance for the correctness of causal events in four of the five metrics, while the ConceptNet NumberBatch model worked best for the correctness of the causal sequence. These evaluation results provide a criterion for model adoption to adaptively support SRL of causal relations in DLEs.

Practitioner notes

What is already known about this topic

  • Accurate monitoring is a prerequisite for effective self-regulation.
  • Students struggle to accurately monitor their comprehension of causal relations in texts.
  • Completing causal diagrams improves students' monitoring accuracy, but there is room for further improvement.
  • Automatic scoring could be used to provide adaptive support during diagramming.

What this paper adds

  • Comparison of four Dutch word vector models combined with four machine learning classifiers for the automatic scoring of students' causal diagrams.
  • Five performance metrics to evaluate the above solutions.
  • Evaluation of the word vector models for estimating the semantic similarity between student and model answers.

Implications for practice and/or policy

  • High-quality word vector models could (em)power adaptive support during causal diagramming via automatic scoring.
  • The evaluated solutions can be embedded in digital learning environments (DLEs).
  • Criteria for model adoption to adaptively support SRL of causal relations in DLEs.
  • The increased saliency of (in)correct answers via automatic scoring might help to improve students' monitoring accuracy.

Abstract Image

为因果关系的自我调节学习提供适应性支持:评估四种荷兰语词汇向量模型
计算语言模型的进步越来越多地为数字学习环境(DLEs)中的自律学习(SRL)提供自适应支持,如通过自动反馈。然而,这些模型的准确性是教育利益相关者(如政策制定者、研究人员、教师和学习者本身)共同关心的问题。我们比较了四种荷兰语模型(即 spaCy 中型、spaCy 大型、FastText 和 ConceptNet NumberBatch)在中学生从说明性文本中学习因果关系时的准确性,并以因果图的完成为支架。由于机器学习依赖于人类标注的数据才能获得最佳效果,因此我们使用了一个包含 10,193 个学生因果图答案的数据集,该数据集是经过十多年的研究整理而成的,使用图表完成干预来加强学生对其文本理解的监控。我们将语言模型与四种流行的机器学习分类器(即逻辑回归、随机森林、支持向量机和神经网络)结合使用,以评估它们在根据事件的正确性及其顺序(即因果结构)对学生的因果图进行自动评分方面的性能。研究了五个性能指标,即准确度、精确度、召回率、F1 和接收者操作特征曲线下面积(ROC-AUC)。在五个指标中,spaCy 中型模型与神经网络分类器相结合,在因果事件的正确性方面取得了四个指标的最佳性能,而 ConceptNet NumberBatch 模型在因果序列的正确性方面效果最佳。这些评估结果为采用模型自适应地支持DLE中因果关系的SRL提供了标准。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
British Journal of Educational Technology
British Journal of Educational Technology EDUCATION & EDUCATIONAL RESEARCH-
CiteScore
15.60
自引率
4.50%
发文量
111
期刊介绍: BJET is a primary source for academics and professionals in the fields of digital educational and training technology throughout the world. The Journal is published by Wiley on behalf of The British Educational Research Association (BERA). It publishes theoretical perspectives, methodological developments and high quality empirical research that demonstrate whether and how applications of instructional/educational technology systems, networks, tools and resources lead to improvements in formal and non-formal education at all levels, from early years through to higher, technical and vocational education, professional development and corporate training.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信