{"title":"TOWARDS EFFECTIVE ARGUMENTATION: DESIGN AND IMPLEMENTATION OF A GENERATIVE AI-BASED EVALUATION AND FEEDBACK SYSTEM","authors":"Hunkoog Jho, Minsu Ha","doi":"10.33225/jbse/24.23.280","DOIUrl":null,"url":null,"abstract":"This study aimed at examining the performance of generative artificial intelligence to extract argumentation elements from text. Thus, the researchers developed a web-based framework to provide automated assessment and feedback relying on a large language model, ChatGPT. The results produced by ChatGPT were compared to human experts across scientific and non-scientific contexts. The findings revealed marked discrepancies in the performance of AI for extracting argument components, with a significant variance between issues of a scientific nature and those that are not. Higher accuracy was noted in identifying claims, data, and qualifiers, as opposed to rebuttals, backing, and warrants. The study illuminated AI's promise for educational applications but also its shortcomings, such as the increased frequency of erroneous element identification when accuracy was low. This highlights the essential need for more in-depth comparative research on models and the further development of AI to enhance its role in supporting argumentation training.\nKeywords: argumentative writing, artificial intelligence, automated assessment, natural language processing, web architecture","PeriodicalId":46424,"journal":{"name":"Journal of Baltic Science Education","volume":null,"pages":null},"PeriodicalIF":1.1000,"publicationDate":"2024-04-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Baltic Science Education","FirstCategoryId":"95","ListUrlMain":"https://doi.org/10.33225/jbse/24.23.280","RegionNum":4,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"EDUCATION & EDUCATIONAL RESEARCH","Score":null,"Total":0}
引用次数: 0
Abstract
This study aimed at examining the performance of generative artificial intelligence to extract argumentation elements from text. Thus, the researchers developed a web-based framework to provide automated assessment and feedback relying on a large language model, ChatGPT. The results produced by ChatGPT were compared to human experts across scientific and non-scientific contexts. The findings revealed marked discrepancies in the performance of AI for extracting argument components, with a significant variance between issues of a scientific nature and those that are not. Higher accuracy was noted in identifying claims, data, and qualifiers, as opposed to rebuttals, backing, and warrants. The study illuminated AI's promise for educational applications but also its shortcomings, such as the increased frequency of erroneous element identification when accuracy was low. This highlights the essential need for more in-depth comparative research on models and the further development of AI to enhance its role in supporting argumentation training.
Keywords: argumentative writing, artificial intelligence, automated assessment, natural language processing, web architecture