Investigating the effect of transformer encoder architecture to improve the reliability of classroom observation ratings on high-inference discourse.

IF 4.6 2区 心理学 Q1 PSYCHOLOGY, EXPERIMENTAL
Jinnie Shin, Wallace N Pinto, Bowen Wang
{"title":"Investigating the effect of transformer encoder architecture to improve the reliability of classroom observation ratings on high-inference discourse.","authors":"Jinnie Shin, Wallace N Pinto, Bowen Wang","doi":"10.3758/s13428-025-02711-5","DOIUrl":null,"url":null,"abstract":"<p><p>This study investigates the effect of transformer encoder architecture on the classification accuracy of high-inference discourse elements in classroom settings. Recognizing the importance of capturing nuanced interactions between students and teachers, our study explores the performance of different transformer models, focusing particularly on the bi-encoder architecture of S-BERT. We evaluated various embedding strategies, along with different pooling methods, to optimize the bi-encoder model's classification accuracy for discourse elements such as High Uptake and Focusing Question. We compared S-BERT's performance with traditional cross-encoding transformer models such as BERT and RoBERTa. Our results demonstrate that S-BERT, particularly with a batch size of 8, learning rate of 2e-5, and specific embedding strategies, significantly outperforms other baseline models, achieving F1 scores up to 0.826 for High Uptake and 0.908 for Focusing Question. Our findings highlighted the importance of customized vectorization strategies, encompassing individual and interaction features (dot-product and absolute distance), and underscores the need to carefully select pooling methods to enhance performance. Our findings offer valuable insights into the design of transformer models for classroom discourse analysis, contributing to the advancement of NLP methods in educational research.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 7","pages":"186"},"PeriodicalIF":4.6000,"publicationDate":"2025-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Behavior Research Methods","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.3758/s13428-025-02711-5","RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHOLOGY, EXPERIMENTAL","Score":null,"Total":0}
引用次数: 0

Abstract

This study investigates the effect of transformer encoder architecture on the classification accuracy of high-inference discourse elements in classroom settings. Recognizing the importance of capturing nuanced interactions between students and teachers, our study explores the performance of different transformer models, focusing particularly on the bi-encoder architecture of S-BERT. We evaluated various embedding strategies, along with different pooling methods, to optimize the bi-encoder model's classification accuracy for discourse elements such as High Uptake and Focusing Question. We compared S-BERT's performance with traditional cross-encoding transformer models such as BERT and RoBERTa. Our results demonstrate that S-BERT, particularly with a batch size of 8, learning rate of 2e-5, and specific embedding strategies, significantly outperforms other baseline models, achieving F1 scores up to 0.826 for High Uptake and 0.908 for Focusing Question. Our findings highlighted the importance of customized vectorization strategies, encompassing individual and interaction features (dot-product and absolute distance), and underscores the need to carefully select pooling methods to enhance performance. Our findings offer valuable insights into the design of transformer models for classroom discourse analysis, contributing to the advancement of NLP methods in educational research.

探讨变压器编码器架构对提高高推理语篇课堂观察评分可靠性的影响。
本研究探讨了变压器编码器结构对课堂语境中高推理语篇要素分类准确率的影响。认识到捕捉学生和教师之间细微互动的重要性,我们的研究探索了不同变压器模型的性能,特别关注S-BERT的双编码器架构。我们评估了各种嵌入策略以及不同的池化方法,以优化双编码器模型对High Uptake和focused Question等话语元素的分类精度。我们将S-BERT与传统的交叉编码转换模型(如BERT和RoBERTa)的性能进行了比较。我们的研究结果表明,S-BERT,特别是在批量大小为8、学习率为25 -5和特定嵌入策略的情况下,显著优于其他基线模型,在High Uptake和focus Question上的F1得分分别高达0.826和0.908。我们的研究结果强调了定制化矢量化策略的重要性,包括个体和交互特征(点积和绝对距离),并强调了仔细选择池化方法以提高性能的必要性。我们的研究结果为课堂话语分析变压器模型的设计提供了有价值的见解,有助于促进NLP方法在教育研究中的发展。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
10.30
自引率
9.30%
发文量
266
期刊介绍: Behavior Research Methods publishes articles concerned with the methods, techniques, and instrumentation of research in experimental psychology. The journal focuses particularly on the use of computer technology in psychological research. An annual special issue is devoted to this field.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信