{"title":"Research on A Machine Scoring Method of Role-Play Section in English Oral Test","authors":"Xinguang Li, Zhihe Yang, Shuai Chen, Shanxian Ma","doi":"10.1145/3498851.3498983","DOIUrl":null,"url":null,"abstract":"Computer-assisted instruction has been widely implemented in language learning since the remarkable development of speech technology and natural language processing. In English teaching, the traditional manual evaluation mode has been replaced by the computer. In this paper, we introduce a machine scoring method of Role-Play Section in the English oral test. Specifically, we design a scoring method of Role-Play section in Computer-based English Listening and Speaking Test (CELST) of the Guangdong college entrance examination (GDCEE), which can simulate human scoring. The role-play section is also called Three Questions (TQ) and Five Answers (FA), so two scoring modules are established, respectively. According to the features of the role-play section and the given manual scoring scale, the scoring task can be regarded as a short text similarity evaluation task, and we propose a corresponding multi-index text similarity evaluation method. Statistic-based word matching and keyword-focused semantic similarity evaluations are adopted in the TQ and FA scoring modules. Additionally, we consider the grammar factor and utilize semantic similarity-combined dependency parsing evaluation for TQ scoring. Based on the above five evaluative indicators, we linearly integrate them and corresponding expertise-based weights with linear regression optimization, thus constructing a comprehensive scoring model of the role-play Section in CELST. Experiments were conducted using real oral data of examinees in GDCEE. Results have indicated that the machine scoring method achieves impressive data consistency with human scoring.","PeriodicalId":89230,"journal":{"name":"Proceedings. IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology","volume":"29 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2021-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings. IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3498851.3498983","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Computer-assisted instruction has been widely implemented in language learning since the remarkable development of speech technology and natural language processing. In English teaching, the traditional manual evaluation mode has been replaced by the computer. In this paper, we introduce a machine scoring method of Role-Play Section in the English oral test. Specifically, we design a scoring method of Role-Play section in Computer-based English Listening and Speaking Test (CELST) of the Guangdong college entrance examination (GDCEE), which can simulate human scoring. The role-play section is also called Three Questions (TQ) and Five Answers (FA), so two scoring modules are established, respectively. According to the features of the role-play section and the given manual scoring scale, the scoring task can be regarded as a short text similarity evaluation task, and we propose a corresponding multi-index text similarity evaluation method. Statistic-based word matching and keyword-focused semantic similarity evaluations are adopted in the TQ and FA scoring modules. Additionally, we consider the grammar factor and utilize semantic similarity-combined dependency parsing evaluation for TQ scoring. Based on the above five evaluative indicators, we linearly integrate them and corresponding expertise-based weights with linear regression optimization, thus constructing a comprehensive scoring model of the role-play Section in CELST. Experiments were conducted using real oral data of examinees in GDCEE. Results have indicated that the machine scoring method achieves impressive data consistency with human scoring.