Genre-based fine-tuning of large language models with self-organizing maps for automated writing evaluation

Stephanie Link, Robert Redmon, Martin Hagan
{"title":"Genre-based fine-tuning of large language models with self-organizing maps for automated writing evaluation","authors":"Stephanie Link,&nbsp;Robert Redmon,&nbsp;Martin Hagan","doi":"10.1016/j.rmal.2025.100219","DOIUrl":null,"url":null,"abstract":"<div><div>Automated Writing Evaluation (AWE) systems have significantly advanced in providing feedback for academic essay writing. However, their predominant focus on sentence-level features highlights the need for a broader approach to AWE development. While genre-based AWE systems aim to address the socio-rhetorical complexities of writing for specific audiences and purposes, their availability remains limited. This scarcity is largely due to methodological constraints in developing robust feedback engines that effectively support discipline-specific writing needs. This article describes a new method for fine-tuning large-language models (LLM) and evaluating model performance, which we refer to as G-FiT Mapping (Genre-based FIne-Tuning with self-organizing maps). This method utilizes semi-automated annotation of genre-based functional-rhetorical units of text to efficiently fine-tune an LLM and then uses self-organizing maps to evaluate and improve network performance. The G-FiT Mapping method resulted in a new automated feedback engine for an intelligent tutoring system called Dissemity, for DISSeminating research with clarITY, that supports discipline-specific, scientific writers in writing for publication. We demonstrate use of G-Fit Mapping for establishing measurable improvements in network performance, offering implications for network interpretation, genre-based AWE, and AI-based learning systems development.</div></div>","PeriodicalId":101075,"journal":{"name":"Research Methods in Applied Linguistics","volume":"4 3","pages":"Article 100219"},"PeriodicalIF":0.0000,"publicationDate":"2025-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Research Methods in Applied Linguistics","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2772766125000400","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Automated Writing Evaluation (AWE) systems have significantly advanced in providing feedback for academic essay writing. However, their predominant focus on sentence-level features highlights the need for a broader approach to AWE development. While genre-based AWE systems aim to address the socio-rhetorical complexities of writing for specific audiences and purposes, their availability remains limited. This scarcity is largely due to methodological constraints in developing robust feedback engines that effectively support discipline-specific writing needs. This article describes a new method for fine-tuning large-language models (LLM) and evaluating model performance, which we refer to as G-FiT Mapping (Genre-based FIne-Tuning with self-organizing maps). This method utilizes semi-automated annotation of genre-based functional-rhetorical units of text to efficiently fine-tune an LLM and then uses self-organizing maps to evaluate and improve network performance. The G-FiT Mapping method resulted in a new automated feedback engine for an intelligent tutoring system called Dissemity, for DISSeminating research with clarITY, that supports discipline-specific, scientific writers in writing for publication. We demonstrate use of G-Fit Mapping for establishing measurable improvements in network performance, offering implications for network interpretation, genre-based AWE, and AI-based learning systems development.
基于体裁的大型语言模型微调与自动写作评估的自组织地图
自动写作评估(AWE)系统在为学术论文写作提供反馈方面取得了重大进展。然而,他们对句子级功能的主要关注突出了对AWE开发的更广泛方法的需求。虽然基于体裁的AWE系统旨在解决特定受众和目的的社会修辞复杂性,但它们的可用性仍然有限。这种稀缺性很大程度上是由于在开发有效支持特定学科写作需求的健壮反馈引擎方面的方法限制。本文描述了一种微调大语言模型(LLM)和评估模型性能的新方法,我们将其称为G-FiT Mapping(基于类型的自组织映射的微调)。该方法利用基于体裁的文本功能修辞单元的半自动注释来有效地微调LLM,然后使用自组织映射来评估和改进网络性能。G-FiT Mapping方法为一个名为“传播”的智能辅导系统提供了一个新的自动反馈引擎,该系统用于传播具有清晰度的研究,支持特定学科的科学作者撰写出版物。我们展示了使用G-Fit Mapping来建立可衡量的网络性能改进,为网络解释、基于类型的AWE和基于人工智能的学习系统开发提供了启示。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
4.10
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信