Parameter-Efficiently Fine-Tuning Large Language Models for Classroom Dialogue Analysis

IF 4.9 3区 教育学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS
Deliang Wang;Yaqian Zheng;Jinjiang Li;Gaowei Chen
{"title":"Parameter-Efficiently Fine-Tuning Large Language Models for Classroom Dialogue Analysis","authors":"Deliang Wang;Yaqian Zheng;Jinjiang Li;Gaowei Chen","doi":"10.1109/TLT.2025.3567995","DOIUrl":null,"url":null,"abstract":"Researchers have increasingly utilized artificial intelligence to automatically analyze classroom dialogue, aiming to provide timely feedback to teachers due to its educational significance. However, traditional machine learning and deep learning models face challenges, such as limited performance and lack of generalizability, across various dimensions of classroom dialogue and educational contexts. Recent efforts to utilize large language models (LLMs) for classroom dialogue analysis have predominantly relied on prompt engineering techniques, primarily due to the high costs associated with full fine-tuning, which has resulted in suboptimal performance and areas needing improvement. We, therefore, propose the application of parameter-efficient fine-tuning (PEFT) techniques to enhance the performance of LLMs in classroom dialogue analysis. Specifically, we utilized low-rank adaptation, a prominent PEFT technique, to fine-tune three state-of-the-art LLMs—Llama-3.2-3B, Gemma-2-9B, and Mistral-7B-v0.3—targeting the analysis of both teachers' and students' dialogic moves within K-12 mathematics lessons. The experimental results indicate that, in comparison to fully fine-tuning BERT and RoBERTa models and prompting LLMs, LLMs fine-tuned using the PEFT technique achieve superior performance. Moreover, the PEFT approach significantly reduced the number of trainable parameters within the LLMs by over 300 times and decreased their training duration. Although the training time for PEFT-tuned LLMs was still longer than that required for fully fine-tuning BERT and RoBERTa, these LLMs demonstrated specialization in this specific dimension and generalizability to other tasks and contexts. We believe that the use of PEFT techniques presents a promising direction for future research in classroom dialogue analysis.","PeriodicalId":49191,"journal":{"name":"IEEE Transactions on Learning Technologies","volume":"18 ","pages":"542-555"},"PeriodicalIF":4.9000,"publicationDate":"2025-03-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Learning Technologies","FirstCategoryId":"95","ListUrlMain":"https://ieeexplore.ieee.org/document/10992249/","RegionNum":3,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0

Abstract

Researchers have increasingly utilized artificial intelligence to automatically analyze classroom dialogue, aiming to provide timely feedback to teachers due to its educational significance. However, traditional machine learning and deep learning models face challenges, such as limited performance and lack of generalizability, across various dimensions of classroom dialogue and educational contexts. Recent efforts to utilize large language models (LLMs) for classroom dialogue analysis have predominantly relied on prompt engineering techniques, primarily due to the high costs associated with full fine-tuning, which has resulted in suboptimal performance and areas needing improvement. We, therefore, propose the application of parameter-efficient fine-tuning (PEFT) techniques to enhance the performance of LLMs in classroom dialogue analysis. Specifically, we utilized low-rank adaptation, a prominent PEFT technique, to fine-tune three state-of-the-art LLMs—Llama-3.2-3B, Gemma-2-9B, and Mistral-7B-v0.3—targeting the analysis of both teachers' and students' dialogic moves within K-12 mathematics lessons. The experimental results indicate that, in comparison to fully fine-tuning BERT and RoBERTa models and prompting LLMs, LLMs fine-tuned using the PEFT technique achieve superior performance. Moreover, the PEFT approach significantly reduced the number of trainable parameters within the LLMs by over 300 times and decreased their training duration. Although the training time for PEFT-tuned LLMs was still longer than that required for fully fine-tuning BERT and RoBERTa, these LLMs demonstrated specialization in this specific dimension and generalizability to other tasks and contexts. We believe that the use of PEFT techniques presents a promising direction for future research in classroom dialogue analysis.
参数有效微调课堂对话分析的大型语言模型
研究人员越来越多地利用人工智能来自动分析课堂对话,旨在为教师提供及时的反馈,因为它具有教育意义。然而,传统的机器学习和深度学习模型面临着挑战,例如在课堂对话和教育背景的各个维度上,性能有限,缺乏通用性。最近利用大型语言模型(llm)进行课堂对话分析的努力主要依赖于快速工程技术,这主要是由于与完全微调相关的高成本,这导致了次优性能和需要改进的领域。因此,我们建议应用参数有效微调(PEFT)技术来提高法学硕士在课堂对话分析中的表现。具体来说,我们利用低阶适应(一种突出的PEFT技术)来微调三个最先进的llms——llama -3.2- 3b、Gemma-2-9B和mistral - 7b -v0.3——针对K-12数学课中教师和学生的对话动作进行分析。实验结果表明,与完全微调BERT和RoBERTa模型和提示llm相比,使用PEFT技术微调的llm具有更好的性能。此外,PEFT方法显着减少了llm中可训练参数的数量超过300倍,并缩短了它们的训练时间。尽管peft调优的法学硕士的培训时间仍然比完全微调BERT和RoBERTa所需的时间长,但这些法学硕士在这一特定维度上表现出专业化,并可推广到其他任务和上下文。我们相信,PEFT技术的使用为课堂对话分析的未来研究提供了一个有希望的方向。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
IEEE Transactions on Learning Technologies
IEEE Transactions on Learning Technologies COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS-
CiteScore
7.50
自引率
5.40%
发文量
82
审稿时长
>12 weeks
期刊介绍: The IEEE Transactions on Learning Technologies covers all advances in learning technologies and their applications, including but not limited to the following topics: innovative online learning systems; intelligent tutors; educational games; simulation systems for education and training; collaborative learning tools; learning with mobile devices; wearable devices and interfaces for learning; personalized and adaptive learning systems; tools for formative and summative assessment; tools for learning analytics and educational data mining; ontologies for learning systems; standards and web services that support learning; authoring tools for learning materials; computer support for peer tutoring; learning via computer-mediated inquiry, field, and lab work; social learning techniques; social networks and infrastructures for learning and knowledge sharing; and creation and management of learning objects.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信