H. Huang, S. Hung, Hsiu-Yi Chao, Jyun-Hong Chen, Tsui-Peng Lin, Ching-Lin Shih
{"title":"Developing and Validating a Computerized Adaptive Testing System for Measuring the English Proficiency of Taiwanese EFL University Students","authors":"H. Huang, S. Hung, Hsiu-Yi Chao, Jyun-Hong Chen, Tsui-Peng Lin, Ching-Lin Shih","doi":"10.1080/15434303.2021.1984490","DOIUrl":"https://doi.org/10.1080/15434303.2021.1984490","url":null,"abstract":"ABSTRACT Prompted by Taiwanese university students’ increasing demand for English proficiency assessment, the absence of a test designed specifically for this demographic subgroup, and the lack of a localized and freely-accessible proficiency measure, this project set out to develop and validate a computerized adaptive English proficiency testing (E-CAT) system for Taiwanese EFL university students. Drawing on the guidelines posited by L2 testing specialists, we devised and followed a six-stage procedure to develop this E-CAT system: determining the test purpose, defining the construct, developing the test items, designing the administration processes, performing the field testing, and constructing the E-CAT system. Upon its completion, we performed two validation studies on the simulated data and as such offered the backing for the generalization inference and the explanation inference, which combined to lend support for the validity argument for the E-CAT score-based interpretations and uses. This project highlighted the importance of test localization, foregrounded the utility of computerized adaptive testing and item response theory in language test development and validation, and generated a localized, free-of-charge English proficiency test for Taiwanese university students to satisfy the graduation benchmark requirement and/or to demonstrate their English proficiency levels when job hunting.","PeriodicalId":46873,"journal":{"name":"Language Assessment Quarterly","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2021-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44445959","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. Stewart, Joseph P. Vitta, Christopher Nicklin, Stuart Mclean, Geoffrey G. Pinchbeck, Brandon Kramer
{"title":"The Relationship between Word Difficulty and Frequency: A Response to Hashimoto (2021)","authors":"J. Stewart, Joseph P. Vitta, Christopher Nicklin, Stuart Mclean, Geoffrey G. Pinchbeck, Brandon Kramer","doi":"10.1080/15434303.2021.1992629","DOIUrl":"https://doi.org/10.1080/15434303.2021.1992629","url":null,"abstract":"ABSTRACT Hashimoto (2021) reported a correlation of −.50 (r 2 = .25) between word frequency rank and difficulty, concluding the construct of modern vocabulary size tests is questionable. In this response we show that the relationship between frequency and difficulty is clear albeit non-linear and demonstrate that if a wider range of frequencies is tested and log transformations are applied, the correlation can approach .80. Finally, while we acknowledge the great promise of knowledge-based word lists, we note that a strong correlation between difficulty and frequency is not, in fact, the primary reason size tests are organized by frequency.","PeriodicalId":46873,"journal":{"name":"Language Assessment Quarterly","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2021-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49354378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Language Testing for Migrants: Co-Constructing Validation","authors":"B. O’Sullivan, Micheline B. Chalhoub-Deville","doi":"10.1080/15434303.2021.1986513","DOIUrl":"https://doi.org/10.1080/15434303.2021.1986513","url":null,"abstract":"The present issue provides a much-needed space to key issues not visible in our discourse in language testing and research. The various articles delve into research, policy, test development, and validity considerations for migrants who are increasingly mandated to take language and literacy tests. The papers point to issues of “test misuse,” bias, negative impact, and altogether different test taker populations, which tend to have low literacy in their first languages. While many of the concerns raised in this special issue relate specifically to the testing of language learners with low print literacy, there are lessons here for test development and validation theory across the board. In our commentary, we will focus primarily on issues of validation. This seems to be a critical theme in all the papers included in the present issue. We and the authors in this special issue argue that the language testing community needs to revisit validity theory considering the intricate connections between language testing and migration policies. Validation, as clearly shown in this issues, needs to be co-constructed by key stakeholder groups at the design, development, administration, research, and use levels.","PeriodicalId":46873,"journal":{"name":"Language Assessment Quarterly","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2021-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43165517","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Putting Low Alphabetic Print Literacy on the Language Testing Agenda","authors":"B. Deygers","doi":"10.1080/15434303.2021.1986512","DOIUrl":"https://doi.org/10.1080/15434303.2021.1986512","url":null,"abstract":"ABSTRACT To date, language testing research has devoted little attention to adult L2 learners with low levels of alphabetic print literacy (LESLLA), even though this population makes up for a substantial proportion of the candidature of language tests used for migration purposes. This special issue focuses on LESLLA learners, shows how literacy impacts test performance beyond what might be considered construct relevant, and how a focus on LESLLA learners invites a reconceptualization of commonly held notions of literacy, test misuse and even validation.","PeriodicalId":46873,"journal":{"name":"Language Assessment Quarterly","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2021-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45042361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Interplay of Text and Image on the Meaning-making Processes of Adult L2 Learners with Emerging Literacy: Implications for Test Design and Evaluation Frameworks","authors":"Jenna A. Altherr Flores","doi":"10.1080/15434303.2021.1984491","DOIUrl":"https://doi.org/10.1080/15434303.2021.1984491","url":null,"abstract":"ABSTRACT This study investigates meaning-making processes in language and literacy assessments. Using a social semiotic perspective, it examines how adult second language learners with emerging literacy self-articulate their understanding of multimodal elements and components utilized in low-stakes assessments, and the strategies they use to make meaning in such tests. The study examines how the intended meaning of assessment prompts aligns with the perceived meaning on the part of participants. To do so, we used an existing assessment tool developed for refugee-background adults in the United States as well as two experimental assessment texts created as part of the research. Data was collected via semi-structured interviews with 14 participants. Interviews were analyzed using a critical multimodal social semiotic approach, informed by systemic functional linguistics. Data was coded according to metafunction and theme, with results organized by test genre element and multimodal component. Results exposed tensions between participants’ responses to textual and visual prompts and the expectations of test designers. Findings reveal that textual composition and assessment practices may be inadvertently biased against individuals with diverging literacy profiles. Results have implications for assessment of this population such that language, literacy, and test socialization should be considered during test design and evaluation framework development.","PeriodicalId":46873,"journal":{"name":"Language Assessment Quarterly","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2021-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45892883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Revisit of Zumbo’s Third Generation DIF: How Are We Doing in Language Testing?","authors":"Hongli Li, C. Hunter, Jacquelyn A. Bialo","doi":"10.1080/15434303.2021.1963253","DOIUrl":"https://doi.org/10.1080/15434303.2021.1963253","url":null,"abstract":"ABSTRACT The purpose of this study is to review the status of differential item functioning (DIF) research in language testing, particularly as it relates to the investigation of sources (or causes) of DIF, which is a defining characteristic of the third generation DIF. This review included 110 DIF studies of language tests dated from 1985 to 2019. We found that DIF researchers did not address sources of DIF more frequently in recent years than in earlier years. Nevertheless, DIF research in language testing has expanded with new DIF analysis procedures, more grouping variables, and more diversified methods for investigating sources of DIF. In addition, in the early years of DIF research, methods to identify sources of DIF relied heavily on content analysis. This review showed that while more sophisticated statistical procedures have been adopted in recent years to address sources of DIF, understanding sources of DIF still remains a challenging task. We also discuss the pros and cons of existing methods to detect sources of DIF and implications for future investigations.","PeriodicalId":46873,"journal":{"name":"Language Assessment Quarterly","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2021-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45640106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Handbook of Automated Scoring: Theory into Practice","authors":"Shaoyan Zou, Jun Wang","doi":"10.1080/15434303.2021.1970168","DOIUrl":"https://doi.org/10.1080/15434303.2021.1970168","url":null,"abstract":"","PeriodicalId":46873,"journal":{"name":"Language Assessment Quarterly","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2021-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42911069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Evolution of Assessment in English Pronunciation: The Case of Hong Kong (1978-2018)","authors":"J. Chan","doi":"10.1080/15434303.2021.1935960","DOIUrl":"https://doi.org/10.1080/15434303.2021.1935960","url":null,"abstract":"ABSTRACT This study tracked the development of Hong Kong’s assessment practices for English pronunciation over the past four decades, with reference to the nativeness and intelligibility principles in L2 pronunciation research and pedagogy. Specifically, it evaluated changes in assessors’ comments on candidates’ English pronunciation performance in school-exit public examinations between 1978 and 2018. Qualitative and quantitative content analyses were conducted on the examination report for each year to identify themes related to candidates’ pronunciation ‘problems’, including ‘word-based’ features (word pronunciation, word stress, segmentals), ‘discourse-based’ features (suprasegmentals) and ‘delivery’ (clarity, fluency, loudness, naturalness, pacing). In the examination reports, candidates’ problems with word-based features (particularly word pronunciation) received the most attention across the decades. Most of the comments in later reports were aligned with the intelligibility principle, particularly at the segmental level (e.g., missing consonants, simplification of consonant clusters, word pronunciation). These assessment practices were potentially influenced by the teaching methods recommended in the different ELT curricula over time (i.e., from an oral-structural to a communicative/task-based language teaching approach), and also by the assessors’ judgements. The paper concludes by proposing a research agenda for the promotion of an evidence-based approach that can inform future assessment practices.","PeriodicalId":46873,"journal":{"name":"Language Assessment Quarterly","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2021-08-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44006162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"At the Intersection of Language Testing Policy, Practice, and Research: An Interview with Yan Jin","authors":"Jason Fan, K. Frost","doi":"10.1080/15434303.2021.1938570","DOIUrl":"https://doi.org/10.1080/15434303.2021.1938570","url":null,"abstract":"","PeriodicalId":46873,"journal":{"name":"Language Assessment Quarterly","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2021-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45431194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Language Test Misuse","authors":"C. Carlsen, L. Rocca","doi":"10.1080/15434303.2021.1947288","DOIUrl":"https://doi.org/10.1080/15434303.2021.1947288","url":null,"abstract":"ABSTRACT During the past two decades, an increasing number of European countries have introduced language requirements for residency, citizenship, and sometimes even for entry to the country and family reunification. As a result, democratic rights as well as basic human rights have come to depend upon an individual’s ability to obtain a certain score on a language test and the consequences of failing may be detrimental. In the field of language testing, this use of language tests is often referred to as test misuse, yet the term lacks a precise definition in the literature. In this paper we propose a definition of test misuse in relation to language tests for migration purposes and focus particular attention on low-literate adult migrants for whom the requirements pose a considerable barrier. The main purpose of this paper is to address the question why language tests are being misused in migration policies, exploring linguistic, political as well as test theoretical explanations. We suggest that a more central role of test misuse in validity theory is essential in order to remedy its lack of research focus in our field.","PeriodicalId":46873,"journal":{"name":"Language Assessment Quarterly","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2021-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"59952583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"文学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}