{"title":"人工与自动机器翻译评价:汉葡翻译质量的灰色关联分析","authors":"Yuqi Sun, Lap-Man Hoi, S. Im","doi":"10.1109/CCAI57533.2023.10201322","DOIUrl":null,"url":null,"abstract":"This study investigates the relationship between manual and automatic machine translation evaluation methodologies by employing Grey Correlation Analysis (GRA) to assess the correlation between Chinese-Portuguese machine translation outputs’ BLEU scores and human evaluation scores based on a proposed evaluation index system. The research aims to provide insights into the factors impacting machine translation quality and the most relevant linguistic dimensions in translation evaluation. The findings reveal that “usability” and “adequacy” exhibit the highest correlation with BLEU scores, while “Semantics” ranks highest among the ten manual evaluation indicators, when correlated with the aggregated human evaluation results, followed by “Correction” (correct information) and “Omission and/or Addition”. The findings contribute to the field of machine translation evaluation by illuminating the complex relationship between manual and automatic evaluation techniques and guiding future improvements in machine translation systems and evaluation methodologies.","PeriodicalId":285760,"journal":{"name":"2023 IEEE 3rd International Conference on Computer Communication and Artificial Intelligence (CCAI)","volume":"86 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Examining Manual and Automatic MT Evaluation: A Grey Relational Analysis for Chinese-Portuguese Translation Quality\",\"authors\":\"Yuqi Sun, Lap-Man Hoi, S. Im\",\"doi\":\"10.1109/CCAI57533.2023.10201322\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This study investigates the relationship between manual and automatic machine translation evaluation methodologies by employing Grey Correlation Analysis (GRA) to assess the correlation between Chinese-Portuguese machine translation outputs’ BLEU scores and human evaluation scores based on a proposed evaluation index system. The research aims to provide insights into the factors impacting machine translation quality and the most relevant linguistic dimensions in translation evaluation. The findings reveal that “usability” and “adequacy” exhibit the highest correlation with BLEU scores, while “Semantics” ranks highest among the ten manual evaluation indicators, when correlated with the aggregated human evaluation results, followed by “Correction” (correct information) and “Omission and/or Addition”. The findings contribute to the field of machine translation evaluation by illuminating the complex relationship between manual and automatic evaluation techniques and guiding future improvements in machine translation systems and evaluation methodologies.\",\"PeriodicalId\":285760,\"journal\":{\"name\":\"2023 IEEE 3rd International Conference on Computer Communication and Artificial Intelligence (CCAI)\",\"volume\":\"86 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-05-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 IEEE 3rd International Conference on Computer Communication and Artificial Intelligence (CCAI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CCAI57533.2023.10201322\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 3rd International Conference on Computer Communication and Artificial Intelligence (CCAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCAI57533.2023.10201322","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Examining Manual and Automatic MT Evaluation: A Grey Relational Analysis for Chinese-Portuguese Translation Quality
This study investigates the relationship between manual and automatic machine translation evaluation methodologies by employing Grey Correlation Analysis (GRA) to assess the correlation between Chinese-Portuguese machine translation outputs’ BLEU scores and human evaluation scores based on a proposed evaluation index system. The research aims to provide insights into the factors impacting machine translation quality and the most relevant linguistic dimensions in translation evaluation. The findings reveal that “usability” and “adequacy” exhibit the highest correlation with BLEU scores, while “Semantics” ranks highest among the ten manual evaluation indicators, when correlated with the aggregated human evaluation results, followed by “Correction” (correct information) and “Omission and/or Addition”. The findings contribute to the field of machine translation evaluation by illuminating the complex relationship between manual and automatic evaluation techniques and guiding future improvements in machine translation systems and evaluation methodologies.