Héctor J. Pijeira-Díaz, Sophia Braumann, Janneke van de Pol, Tamara van Gog, Anique B. H. de Bruin
{"title":"为因果关系的自我调节学习提供适应性支持:评估四种荷兰语词汇向量模型","authors":"Héctor J. Pijeira-Díaz, Sophia Braumann, Janneke van de Pol, Tamara van Gog, Anique B. H. de Bruin","doi":"10.1111/bjet.13431","DOIUrl":null,"url":null,"abstract":"<div>\n \n <section>\n \n <p>Advances in computational language models increasingly enable adaptive support for self-regulated learning (SRL) in digital learning environments (DLEs; eg, via automated feedback). However, the accuracy of those models is a common concern for educational stakeholders (eg, policymakers, researchers, teachers and learners themselves). We compared the accuracy of four Dutch language models (ie, spaCy medium, spaCy large, FastText and ConceptNet NumberBatch) in the context of secondary school students' learning of causal relations from expository texts, scaffolded by causal diagram completion. Since machine learning relies on human-labelled data for the best results, we used a dataset with 10,193 students' causal diagram answers, compiled over a decade of research using a diagram completion intervention to enhance students' monitoring of their text comprehension. The language models were used in combination with four popular machine learning classifiers (ie, logistic regression, random forests, support vector machine and neural networks) to evaluate their performance on automatically scoring students' causal diagrams in terms of the correctness of events and their sequence (ie, the causal structure). Five performance metrics were studied, namely accuracy, precision, recall, F<sub>1</sub> and the area under the curve of the receiver operating characteristic (ROC-AUC). The spaCy medium model combined with the neural network classifier achieved the best performance for the correctness of causal events in four of the five metrics, while the ConceptNet NumberBatch model worked best for the correctness of the causal sequence. These evaluation results provide a criterion for model adoption to adaptively support SRL of causal relations in DLEs.</p>\n </section>\n \n <section>\n \n <div>\n \n <div>\n \n <h3>Practitioner notes</h3>\n <p>What is already known about this topic\n\n </p><ul>\n \n <li>Accurate monitoring is a prerequisite for effective self-regulation.</li>\n \n <li>Students struggle to accurately monitor their comprehension of causal relations in texts.</li>\n \n <li>Completing causal diagrams improves students' monitoring accuracy, but there is room for further improvement.</li>\n \n <li>Automatic scoring could be used to provide adaptive support during diagramming.</li>\n </ul>\n <p>What this paper adds\n\n </p><ul>\n \n <li>Comparison of four Dutch word vector models combined with four machine learning classifiers for the automatic scoring of students' causal diagrams.</li>\n \n <li>Five performance metrics to evaluate the above solutions.</li>\n \n <li>Evaluation of the word vector models for estimating the semantic similarity between student and model answers.</li>\n </ul>\n <p>Implications for practice and/or policy\n\n </p><ul>\n \n <li>High-quality word vector models could (em)power adaptive support during causal diagramming via automatic scoring.</li>\n \n <li>The evaluated solutions can be embedded in digital learning environments (DLEs).</li>\n \n <li>Criteria for model adoption to adaptively support SRL of causal relations in DLEs.</li>\n \n <li>The increased saliency of (in)correct answers via automatic scoring might help to improve students' monitoring accuracy.</li>\n </ul>\n </div>\n </div>\n </section>\n </div>","PeriodicalId":48315,"journal":{"name":"British Journal of Educational Technology","volume":"55 4","pages":"1354-1375"},"PeriodicalIF":6.7000,"publicationDate":"2024-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bjet.13431","citationCount":"0","resultStr":"{\"title\":\"Towards adaptive support for self-regulated learning of causal relations: Evaluating four Dutch word vector models\",\"authors\":\"Héctor J. Pijeira-Díaz, Sophia Braumann, Janneke van de Pol, Tamara van Gog, Anique B. H. de Bruin\",\"doi\":\"10.1111/bjet.13431\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n <section>\\n \\n <p>Advances in computational language models increasingly enable adaptive support for self-regulated learning (SRL) in digital learning environments (DLEs; eg, via automated feedback). However, the accuracy of those models is a common concern for educational stakeholders (eg, policymakers, researchers, teachers and learners themselves). We compared the accuracy of four Dutch language models (ie, spaCy medium, spaCy large, FastText and ConceptNet NumberBatch) in the context of secondary school students' learning of causal relations from expository texts, scaffolded by causal diagram completion. Since machine learning relies on human-labelled data for the best results, we used a dataset with 10,193 students' causal diagram answers, compiled over a decade of research using a diagram completion intervention to enhance students' monitoring of their text comprehension. The language models were used in combination with four popular machine learning classifiers (ie, logistic regression, random forests, support vector machine and neural networks) to evaluate their performance on automatically scoring students' causal diagrams in terms of the correctness of events and their sequence (ie, the causal structure). Five performance metrics were studied, namely accuracy, precision, recall, F<sub>1</sub> and the area under the curve of the receiver operating characteristic (ROC-AUC). The spaCy medium model combined with the neural network classifier achieved the best performance for the correctness of causal events in four of the five metrics, while the ConceptNet NumberBatch model worked best for the correctness of the causal sequence. These evaluation results provide a criterion for model adoption to adaptively support SRL of causal relations in DLEs.</p>\\n </section>\\n \\n <section>\\n \\n <div>\\n \\n <div>\\n \\n <h3>Practitioner notes</h3>\\n <p>What is already known about this topic\\n\\n </p><ul>\\n \\n <li>Accurate monitoring is a prerequisite for effective self-regulation.</li>\\n \\n <li>Students struggle to accurately monitor their comprehension of causal relations in texts.</li>\\n \\n <li>Completing causal diagrams improves students' monitoring accuracy, but there is room for further improvement.</li>\\n \\n <li>Automatic scoring could be used to provide adaptive support during diagramming.</li>\\n </ul>\\n <p>What this paper adds\\n\\n </p><ul>\\n \\n <li>Comparison of four Dutch word vector models combined with four machine learning classifiers for the automatic scoring of students' causal diagrams.</li>\\n \\n <li>Five performance metrics to evaluate the above solutions.</li>\\n \\n <li>Evaluation of the word vector models for estimating the semantic similarity between student and model answers.</li>\\n </ul>\\n <p>Implications for practice and/or policy\\n\\n </p><ul>\\n \\n <li>High-quality word vector models could (em)power adaptive support during causal diagramming via automatic scoring.</li>\\n \\n <li>The evaluated solutions can be embedded in digital learning environments (DLEs).</li>\\n \\n <li>Criteria for model adoption to adaptively support SRL of causal relations in DLEs.</li>\\n \\n <li>The increased saliency of (in)correct answers via automatic scoring might help to improve students' monitoring accuracy.</li>\\n </ul>\\n </div>\\n </div>\\n </section>\\n </div>\",\"PeriodicalId\":48315,\"journal\":{\"name\":\"British Journal of Educational Technology\",\"volume\":\"55 4\",\"pages\":\"1354-1375\"},\"PeriodicalIF\":6.7000,\"publicationDate\":\"2024-01-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bjet.13431\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"British Journal of Educational Technology\",\"FirstCategoryId\":\"95\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1111/bjet.13431\",\"RegionNum\":1,\"RegionCategory\":\"教育学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"EDUCATION & EDUCATIONAL RESEARCH\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"British Journal of Educational Technology","FirstCategoryId":"95","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/bjet.13431","RegionNum":1,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"EDUCATION & EDUCATIONAL RESEARCH","Score":null,"Total":0}
Towards adaptive support for self-regulated learning of causal relations: Evaluating four Dutch word vector models
Advances in computational language models increasingly enable adaptive support for self-regulated learning (SRL) in digital learning environments (DLEs; eg, via automated feedback). However, the accuracy of those models is a common concern for educational stakeholders (eg, policymakers, researchers, teachers and learners themselves). We compared the accuracy of four Dutch language models (ie, spaCy medium, spaCy large, FastText and ConceptNet NumberBatch) in the context of secondary school students' learning of causal relations from expository texts, scaffolded by causal diagram completion. Since machine learning relies on human-labelled data for the best results, we used a dataset with 10,193 students' causal diagram answers, compiled over a decade of research using a diagram completion intervention to enhance students' monitoring of their text comprehension. The language models were used in combination with four popular machine learning classifiers (ie, logistic regression, random forests, support vector machine and neural networks) to evaluate their performance on automatically scoring students' causal diagrams in terms of the correctness of events and their sequence (ie, the causal structure). Five performance metrics were studied, namely accuracy, precision, recall, F1 and the area under the curve of the receiver operating characteristic (ROC-AUC). The spaCy medium model combined with the neural network classifier achieved the best performance for the correctness of causal events in four of the five metrics, while the ConceptNet NumberBatch model worked best for the correctness of the causal sequence. These evaluation results provide a criterion for model adoption to adaptively support SRL of causal relations in DLEs.
Practitioner notes
What is already known about this topic
Accurate monitoring is a prerequisite for effective self-regulation.
Students struggle to accurately monitor their comprehension of causal relations in texts.
Completing causal diagrams improves students' monitoring accuracy, but there is room for further improvement.
Automatic scoring could be used to provide adaptive support during diagramming.
What this paper adds
Comparison of four Dutch word vector models combined with four machine learning classifiers for the automatic scoring of students' causal diagrams.
Five performance metrics to evaluate the above solutions.
Evaluation of the word vector models for estimating the semantic similarity between student and model answers.
Implications for practice and/or policy
High-quality word vector models could (em)power adaptive support during causal diagramming via automatic scoring.
The evaluated solutions can be embedded in digital learning environments (DLEs).
Criteria for model adoption to adaptively support SRL of causal relations in DLEs.
The increased saliency of (in)correct answers via automatic scoring might help to improve students' monitoring accuracy.
期刊介绍:
BJET is a primary source for academics and professionals in the fields of digital educational and training technology throughout the world. The Journal is published by Wiley on behalf of The British Educational Research Association (BERA). It publishes theoretical perspectives, methodological developments and high quality empirical research that demonstrate whether and how applications of instructional/educational technology systems, networks, tools and resources lead to improvements in formal and non-formal education at all levels, from early years through to higher, technical and vocational education, professional development and corporate training.