Ran Song , Shengxiang Gao , Xiaofei Gao , Cunli Mao , Zhengtao Yu
{"title":"MKE-PLLM:基于预训练大语言模型的多语言知识编辑基准","authors":"Ran Song , Shengxiang Gao , Xiaofei Gao , Cunli Mao , Zhengtao Yu","doi":"10.1016/j.neucom.2025.130979","DOIUrl":null,"url":null,"abstract":"<div><div>Multilingual large language models have demonstrated remarkable performance across various downstream tasks but are still plagued by factuality errors. Knowledge editing aims to correct these errors by modifying the internal knowledge of pre-trained models. However, current knowledge editing methods primarily focus on monolingual settings, neglecting the complexities and interdependencies within multilingual scenarios. Furthermore, benchmarks specifically designed for multilingual knowledge editing are relatively scarce. Addressing this gap, this paper constructs a novel multilingual knowledge editing benchmark. This benchmark comprehensively evaluates methods for mLLMs based on accuracy, reliability, generalization, and consistency. To ensure the robustness and usability of the benchmark, we conducted detailed analysis and validation. Concurrently, we propose a baseline method that adapts existing monolingual knowledge editing techniques to the multilingual environment. Extensive experimental results demonstrate the effectiveness of our constructed benchmark in evaluating multilingual knowledge editing.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"651 ","pages":"Article 130979"},"PeriodicalIF":5.5000,"publicationDate":"2025-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"MKE-PLLM: A benchmark for multilingual knowledge editing on pretrained large language model\",\"authors\":\"Ran Song , Shengxiang Gao , Xiaofei Gao , Cunli Mao , Zhengtao Yu\",\"doi\":\"10.1016/j.neucom.2025.130979\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Multilingual large language models have demonstrated remarkable performance across various downstream tasks but are still plagued by factuality errors. Knowledge editing aims to correct these errors by modifying the internal knowledge of pre-trained models. However, current knowledge editing methods primarily focus on monolingual settings, neglecting the complexities and interdependencies within multilingual scenarios. Furthermore, benchmarks specifically designed for multilingual knowledge editing are relatively scarce. Addressing this gap, this paper constructs a novel multilingual knowledge editing benchmark. This benchmark comprehensively evaluates methods for mLLMs based on accuracy, reliability, generalization, and consistency. To ensure the robustness and usability of the benchmark, we conducted detailed analysis and validation. Concurrently, we propose a baseline method that adapts existing monolingual knowledge editing techniques to the multilingual environment. Extensive experimental results demonstrate the effectiveness of our constructed benchmark in evaluating multilingual knowledge editing.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"651 \",\"pages\":\"Article 130979\"},\"PeriodicalIF\":5.5000,\"publicationDate\":\"2025-07-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225016510\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225016510","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
MKE-PLLM: A benchmark for multilingual knowledge editing on pretrained large language model
Multilingual large language models have demonstrated remarkable performance across various downstream tasks but are still plagued by factuality errors. Knowledge editing aims to correct these errors by modifying the internal knowledge of pre-trained models. However, current knowledge editing methods primarily focus on monolingual settings, neglecting the complexities and interdependencies within multilingual scenarios. Furthermore, benchmarks specifically designed for multilingual knowledge editing are relatively scarce. Addressing this gap, this paper constructs a novel multilingual knowledge editing benchmark. This benchmark comprehensively evaluates methods for mLLMs based on accuracy, reliability, generalization, and consistency. To ensure the robustness and usability of the benchmark, we conducted detailed analysis and validation. Concurrently, we propose a baseline method that adapts existing monolingual knowledge editing techniques to the multilingual environment. Extensive experimental results demonstrate the effectiveness of our constructed benchmark in evaluating multilingual knowledge editing.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.