{"title":"基于体裁的大型语言模型微调与自动写作评估的自组织地图","authors":"Stephanie Link, Robert Redmon, Martin Hagan","doi":"10.1016/j.rmal.2025.100219","DOIUrl":null,"url":null,"abstract":"<div><div>Automated Writing Evaluation (AWE) systems have significantly advanced in providing feedback for academic essay writing. However, their predominant focus on sentence-level features highlights the need for a broader approach to AWE development. While genre-based AWE systems aim to address the socio-rhetorical complexities of writing for specific audiences and purposes, their availability remains limited. This scarcity is largely due to methodological constraints in developing robust feedback engines that effectively support discipline-specific writing needs. This article describes a new method for fine-tuning large-language models (LLM) and evaluating model performance, which we refer to as G-FiT Mapping (Genre-based FIne-Tuning with self-organizing maps). This method utilizes semi-automated annotation of genre-based functional-rhetorical units of text to efficiently fine-tune an LLM and then uses self-organizing maps to evaluate and improve network performance. The G-FiT Mapping method resulted in a new automated feedback engine for an intelligent tutoring system called Dissemity, for DISSeminating research with clarITY, that supports discipline-specific, scientific writers in writing for publication. We demonstrate use of G-Fit Mapping for establishing measurable improvements in network performance, offering implications for network interpretation, genre-based AWE, and AI-based learning systems development.</div></div>","PeriodicalId":101075,"journal":{"name":"Research Methods in Applied Linguistics","volume":"4 3","pages":"Article 100219"},"PeriodicalIF":0.0000,"publicationDate":"2025-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Genre-based fine-tuning of large language models with self-organizing maps for automated writing evaluation\",\"authors\":\"Stephanie Link, Robert Redmon, Martin Hagan\",\"doi\":\"10.1016/j.rmal.2025.100219\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Automated Writing Evaluation (AWE) systems have significantly advanced in providing feedback for academic essay writing. However, their predominant focus on sentence-level features highlights the need for a broader approach to AWE development. While genre-based AWE systems aim to address the socio-rhetorical complexities of writing for specific audiences and purposes, their availability remains limited. This scarcity is largely due to methodological constraints in developing robust feedback engines that effectively support discipline-specific writing needs. This article describes a new method for fine-tuning large-language models (LLM) and evaluating model performance, which we refer to as G-FiT Mapping (Genre-based FIne-Tuning with self-organizing maps). This method utilizes semi-automated annotation of genre-based functional-rhetorical units of text to efficiently fine-tune an LLM and then uses self-organizing maps to evaluate and improve network performance. The G-FiT Mapping method resulted in a new automated feedback engine for an intelligent tutoring system called Dissemity, for DISSeminating research with clarITY, that supports discipline-specific, scientific writers in writing for publication. We demonstrate use of G-Fit Mapping for establishing measurable improvements in network performance, offering implications for network interpretation, genre-based AWE, and AI-based learning systems development.</div></div>\",\"PeriodicalId\":101075,\"journal\":{\"name\":\"Research Methods in Applied Linguistics\",\"volume\":\"4 3\",\"pages\":\"Article 100219\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-06-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Research Methods in Applied Linguistics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2772766125000400\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Research Methods in Applied Linguistics","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2772766125000400","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Genre-based fine-tuning of large language models with self-organizing maps for automated writing evaluation
Automated Writing Evaluation (AWE) systems have significantly advanced in providing feedback for academic essay writing. However, their predominant focus on sentence-level features highlights the need for a broader approach to AWE development. While genre-based AWE systems aim to address the socio-rhetorical complexities of writing for specific audiences and purposes, their availability remains limited. This scarcity is largely due to methodological constraints in developing robust feedback engines that effectively support discipline-specific writing needs. This article describes a new method for fine-tuning large-language models (LLM) and evaluating model performance, which we refer to as G-FiT Mapping (Genre-based FIne-Tuning with self-organizing maps). This method utilizes semi-automated annotation of genre-based functional-rhetorical units of text to efficiently fine-tune an LLM and then uses self-organizing maps to evaluate and improve network performance. The G-FiT Mapping method resulted in a new automated feedback engine for an intelligent tutoring system called Dissemity, for DISSeminating research with clarITY, that supports discipline-specific, scientific writers in writing for publication. We demonstrate use of G-Fit Mapping for establishing measurable improvements in network performance, offering implications for network interpretation, genre-based AWE, and AI-based learning systems development.