{"title":"用机器学习方法预测英语听力文本的CEFR水平","authors":"Christopher Robert Cooper","doi":"10.1016/j.rmal.2025.100234","DOIUrl":null,"url":null,"abstract":"<div><div>Comprehension in listening texts is often judged by lexical coverage. However, this might not be easily interpretable for language teachers. The CEFR is becoming increasingly influential due to its standardized descriptors across languages. Learners are often placed into classes based on proficiency level, therefore a CEFR level is likely more interpretable than lexical coverage when judging listening text difficulty. Machine learning methods have been used to predict the CEFR level of English reading texts and learner writing, but no such studies exist for listening. The current study hopes to bridge this gap by investigating the potential to predict the CEFR level of listening texts. A corpus of CEFR-labelled listening texts (728 texts, 345,104 words) was compiled for text classification. Three types of variables were created from the corpus data to evaluate comparative predictive accuracy. The first method used linguistic and acoustic features. The others used text embeddings, which represent semantic meaning. The data was split into four classes: A1, A2, B1, and B2+. The accuracy of each method was evaluated by comparing the predicted label in the test data with the label from the original text. The most accurate method used OpenAI embeddings and Support Vector Machines. The overall accuracy was 0.81, with macro averages of precision = 0.75, recall = 0.78, and f-score = 0.76, indicating balanced classification performance across CEFR levels. This method has the potential to predict the CEFR level of listening texts, which could help practitioners and researchers match learners and participants to appropriate listening texts.</div></div>","PeriodicalId":101075,"journal":{"name":"Research Methods in Applied Linguistics","volume":"4 3","pages":"Article 100234"},"PeriodicalIF":0.0000,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Predicting the CEFR level of English listening texts with machine learning methods\",\"authors\":\"Christopher Robert Cooper\",\"doi\":\"10.1016/j.rmal.2025.100234\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Comprehension in listening texts is often judged by lexical coverage. However, this might not be easily interpretable for language teachers. The CEFR is becoming increasingly influential due to its standardized descriptors across languages. Learners are often placed into classes based on proficiency level, therefore a CEFR level is likely more interpretable than lexical coverage when judging listening text difficulty. Machine learning methods have been used to predict the CEFR level of English reading texts and learner writing, but no such studies exist for listening. The current study hopes to bridge this gap by investigating the potential to predict the CEFR level of listening texts. A corpus of CEFR-labelled listening texts (728 texts, 345,104 words) was compiled for text classification. Three types of variables were created from the corpus data to evaluate comparative predictive accuracy. The first method used linguistic and acoustic features. The others used text embeddings, which represent semantic meaning. The data was split into four classes: A1, A2, B1, and B2+. The accuracy of each method was evaluated by comparing the predicted label in the test data with the label from the original text. The most accurate method used OpenAI embeddings and Support Vector Machines. The overall accuracy was 0.81, with macro averages of precision = 0.75, recall = 0.78, and f-score = 0.76, indicating balanced classification performance across CEFR levels. This method has the potential to predict the CEFR level of listening texts, which could help practitioners and researchers match learners and participants to appropriate listening texts.</div></div>\",\"PeriodicalId\":101075,\"journal\":{\"name\":\"Research Methods in Applied Linguistics\",\"volume\":\"4 3\",\"pages\":\"Article 100234\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Research Methods in Applied Linguistics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2772766125000552\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Research Methods in Applied Linguistics","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2772766125000552","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Predicting the CEFR level of English listening texts with machine learning methods
Comprehension in listening texts is often judged by lexical coverage. However, this might not be easily interpretable for language teachers. The CEFR is becoming increasingly influential due to its standardized descriptors across languages. Learners are often placed into classes based on proficiency level, therefore a CEFR level is likely more interpretable than lexical coverage when judging listening text difficulty. Machine learning methods have been used to predict the CEFR level of English reading texts and learner writing, but no such studies exist for listening. The current study hopes to bridge this gap by investigating the potential to predict the CEFR level of listening texts. A corpus of CEFR-labelled listening texts (728 texts, 345,104 words) was compiled for text classification. Three types of variables were created from the corpus data to evaluate comparative predictive accuracy. The first method used linguistic and acoustic features. The others used text embeddings, which represent semantic meaning. The data was split into four classes: A1, A2, B1, and B2+. The accuracy of each method was evaluated by comparing the predicted label in the test data with the label from the original text. The most accurate method used OpenAI embeddings and Support Vector Machines. The overall accuracy was 0.81, with macro averages of precision = 0.75, recall = 0.78, and f-score = 0.76, indicating balanced classification performance across CEFR levels. This method has the potential to predict the CEFR level of listening texts, which could help practitioners and researchers match learners and participants to appropriate listening texts.