{"title":"Testing crowdsourcing as a means of recruitment for the comparative judgement of L2 argumentative essays","authors":"Peter Thwaites , Magali Paquot","doi":"10.1016/j.jslw.2025.101207","DOIUrl":null,"url":null,"abstract":"<div><div>Comparative judgement (CJ) is an assessment method in which a large number of pairwise comparisons between learner productions are used to generate scales ranking each item from strongest to weakest. Recent research has suggested that combining CJ with various approaches to judge recruitment, including community-driven and crowdsourcing methods, holds promise as a method of assessing L2 writing, especially for research purposes. However, the majority of studies to date have tested CJ only using relatively simple, easily evaluated sets of texts. There remains insufficient evidence of the method’s potential for assessing more complex texts, particularly when the comparisons are being conducted by crowdsourced assessors. This study seeks to address this problem by testing the reliability and validity of a crowdsourced form of CJ for the assessment of texts which are longer, more topically diverse, and more homogeneous in proficiency than those used in earlier studies. The results suggest that CJ can be conducted with crowdsourced judges to generate reliable assessments of L2 writing, and provide initial evidence of concurrent validity. However, there are drawbacks in terms of efficiency.</div></div>","PeriodicalId":47934,"journal":{"name":"Journal of Second Language Writing","volume":"68 ","pages":"Article 101207"},"PeriodicalIF":5.0000,"publicationDate":"2025-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Second Language Writing","FirstCategoryId":"98","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1060374325000323","RegionNum":1,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"LINGUISTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Comparative judgement (CJ) is an assessment method in which a large number of pairwise comparisons between learner productions are used to generate scales ranking each item from strongest to weakest. Recent research has suggested that combining CJ with various approaches to judge recruitment, including community-driven and crowdsourcing methods, holds promise as a method of assessing L2 writing, especially for research purposes. However, the majority of studies to date have tested CJ only using relatively simple, easily evaluated sets of texts. There remains insufficient evidence of the method’s potential for assessing more complex texts, particularly when the comparisons are being conducted by crowdsourced assessors. This study seeks to address this problem by testing the reliability and validity of a crowdsourced form of CJ for the assessment of texts which are longer, more topically diverse, and more homogeneous in proficiency than those used in earlier studies. The results suggest that CJ can be conducted with crowdsourced judges to generate reliable assessments of L2 writing, and provide initial evidence of concurrent validity. However, there are drawbacks in terms of efficiency.
期刊介绍:
The Journal of Second Language Writing is devoted to publishing theoretically grounded reports of research and discussions that represent a significant contribution to current understandings of central issues in second and foreign language writing and writing instruction. Some areas of interest are personal characteristics and attitudes of L2 writers, L2 writers'' composing processes, features of L2 writers'' texts, readers'' responses to L2 writing, assessment/evaluation of L2 writing, contexts (cultural, social, political, institutional) for L2 writing, and any other topic clearly relevant to L2 writing theory, research, or instruction.