Wensheng Gan , Zhenyao Ning , Zhenlian Qi , Philip S. Yu
{"title":"专家组合(MoE):大数据视角","authors":"Wensheng Gan , Zhenyao Ning , Zhenlian Qi , Philip S. Yu","doi":"10.1016/j.inffus.2025.103664","DOIUrl":null,"url":null,"abstract":"<div><div>As the era of big data arrives, traditional artificial intelligence algorithms have difficulty processing the demands of massive and diverse data. Mixture of experts (MoE) has shown excellent performance and broad application prospects. This paper provides an in-depth review and analysis of the latest progress in this field from multiple perspectives, including the basic principles, algorithmic models, key technical challenges, and application practices of MoE. First, we introduce the basic concept of MoE and its core idea and elaborate on its advantages over traditional single models. Then, we discuss the basic architecture of MoE and its main components, including the gating network, expert networks, and learning algorithms. Next, we review the applications of MoE in addressing key technical issues in big data, including high-dimensional sparse data modeling, heterogeneous multisource data fusion, real-time online learning, and the interpretability of the model. For each challenge, we provide specific MoE solutions and their innovations. Furthermore, we summarize the typical use cases of MoE in various application domains, including natural language processing, computer vision, and recommendation systems, and analyze their outstanding achievements. This fully demonstrates the powerful capability of MoE in big data processing. We also analyze the advantages of MoE in big data environments, including high scalability, efficient resource utilization, and better generalization ability, as well as the challenges it faces, such as load imbalance and expert utilization, gating network stability, and training difficulty. Finally, we explore the future development trends of MoE, including the improvement of model generalization capabilities, the enhancement of algorithmic interpretability, and the increase in system automation levels. We believe that MoE will become an important paradigm of artificial intelligence in the era of big data. In summary, this paper systematically elaborates on the principles, techniques, and applications of MoE in big data processing, providing theoretical and practical references to further promote the application of MoE in real scenarios.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"127 ","pages":"Article 103664"},"PeriodicalIF":15.5000,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Mixture of experts (MoE): A big data perspective\",\"authors\":\"Wensheng Gan , Zhenyao Ning , Zhenlian Qi , Philip S. Yu\",\"doi\":\"10.1016/j.inffus.2025.103664\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>As the era of big data arrives, traditional artificial intelligence algorithms have difficulty processing the demands of massive and diverse data. Mixture of experts (MoE) has shown excellent performance and broad application prospects. This paper provides an in-depth review and analysis of the latest progress in this field from multiple perspectives, including the basic principles, algorithmic models, key technical challenges, and application practices of MoE. First, we introduce the basic concept of MoE and its core idea and elaborate on its advantages over traditional single models. Then, we discuss the basic architecture of MoE and its main components, including the gating network, expert networks, and learning algorithms. Next, we review the applications of MoE in addressing key technical issues in big data, including high-dimensional sparse data modeling, heterogeneous multisource data fusion, real-time online learning, and the interpretability of the model. For each challenge, we provide specific MoE solutions and their innovations. Furthermore, we summarize the typical use cases of MoE in various application domains, including natural language processing, computer vision, and recommendation systems, and analyze their outstanding achievements. This fully demonstrates the powerful capability of MoE in big data processing. We also analyze the advantages of MoE in big data environments, including high scalability, efficient resource utilization, and better generalization ability, as well as the challenges it faces, such as load imbalance and expert utilization, gating network stability, and training difficulty. Finally, we explore the future development trends of MoE, including the improvement of model generalization capabilities, the enhancement of algorithmic interpretability, and the increase in system automation levels. We believe that MoE will become an important paradigm of artificial intelligence in the era of big data. In summary, this paper systematically elaborates on the principles, techniques, and applications of MoE in big data processing, providing theoretical and practical references to further promote the application of MoE in real scenarios.</div></div>\",\"PeriodicalId\":50367,\"journal\":{\"name\":\"Information Fusion\",\"volume\":\"127 \",\"pages\":\"Article 103664\"},\"PeriodicalIF\":15.5000,\"publicationDate\":\"2025-09-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Fusion\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1566253525007365\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525007365","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
As the era of big data arrives, traditional artificial intelligence algorithms have difficulty processing the demands of massive and diverse data. Mixture of experts (MoE) has shown excellent performance and broad application prospects. This paper provides an in-depth review and analysis of the latest progress in this field from multiple perspectives, including the basic principles, algorithmic models, key technical challenges, and application practices of MoE. First, we introduce the basic concept of MoE and its core idea and elaborate on its advantages over traditional single models. Then, we discuss the basic architecture of MoE and its main components, including the gating network, expert networks, and learning algorithms. Next, we review the applications of MoE in addressing key technical issues in big data, including high-dimensional sparse data modeling, heterogeneous multisource data fusion, real-time online learning, and the interpretability of the model. For each challenge, we provide specific MoE solutions and their innovations. Furthermore, we summarize the typical use cases of MoE in various application domains, including natural language processing, computer vision, and recommendation systems, and analyze their outstanding achievements. This fully demonstrates the powerful capability of MoE in big data processing. We also analyze the advantages of MoE in big data environments, including high scalability, efficient resource utilization, and better generalization ability, as well as the challenges it faces, such as load imbalance and expert utilization, gating network stability, and training difficulty. Finally, we explore the future development trends of MoE, including the improvement of model generalization capabilities, the enhancement of algorithmic interpretability, and the increase in system automation levels. We believe that MoE will become an important paradigm of artificial intelligence in the era of big data. In summary, this paper systematically elaborates on the principles, techniques, and applications of MoE in big data processing, providing theoretical and practical references to further promote the application of MoE in real scenarios.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.