N. Herrera, Jorge J. Villalón, Gonzalo Muñoz, Gabriela Baez
{"title":"Aleph: An indicator of decision-making patterns within collaborative marking sessions for writing assessment that can predict reliability","authors":"N. Herrera, Jorge J. Villalón, Gonzalo Muñoz, Gabriela Baez","doi":"10.1109/icalt49669.2020.00041","DOIUrl":null,"url":null,"abstract":"Assessing writing is difficult due to language inherent subjectivity. Rubrics can help tackle this problem, although collaborative training sessions on their interpretation are required in order to reach consensus between markers and ensure minimum reliability levels. This article reports the construction of Aleph; an indicator built on raters’ decision-making behavior in collaborative environments, which can explain agreement between markers. Evaluated within real training sessions of three markers for a high-stake written test, it showed excellent results on identifying markers’ patterns and a good correlation with Fleiss’ Kappa on agreement.","PeriodicalId":153823,"journal":{"name":"2020 IEEE 20th International Conference on Advanced Learning Technologies (ICALT)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE 20th International Conference on Advanced Learning Technologies (ICALT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/icalt49669.2020.00041","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Assessing writing is difficult due to language inherent subjectivity. Rubrics can help tackle this problem, although collaborative training sessions on their interpretation are required in order to reach consensus between markers and ensure minimum reliability levels. This article reports the construction of Aleph; an indicator built on raters’ decision-making behavior in collaborative environments, which can explain agreement between markers. Evaluated within real training sessions of three markers for a high-stake written test, it showed excellent results on identifying markers’ patterns and a good correlation with Fleiss’ Kappa on agreement.