Youcheng Yan , Jinshuo Liu , Juan Deng , Junyan Li , Lina Wang , Jeff Z. Pan
{"title":"协作大型和小型语言模型,实现多模式紧急谣言检测","authors":"Youcheng Yan , Jinshuo Liu , Juan Deng , Junyan Li , Lina Wang , Jeff Z. Pan","doi":"10.1016/j.neunet.2025.107625","DOIUrl":null,"url":null,"abstract":"<div><div>Multi-modal emergency rumors are spreading in the current digital era, causing significant disruptions and negative impacts. Most existing methods focus on exploring rumor detection using individual small language models (SLMs) or large language models (LLMs), achieving a certain degree of success but with underlying issues. Approaches based on SLMs have reached a bottleneck due to their limited knowledge and capacity. In contrast, LLMs have unique strengths in deep analysis that compensate for the weaknesses of SLMs; however, they struggle to select and integrate analyses to draw appropriate conclusions. Furthermore, recent works on multi-modal feature fusion remain superficial, limiting the ability of these models to fully comprehend and identify rumors. In this work, we propose Collaborate Large and Small Language Models for Multi-Modal Emergency Rumor Detection (M2ERD). Specifically, it consists of two main components. First, LLMs generate multi-dimensional rationales based on multi-perspective prompts, from which SLMs selectively derive insights for rumor detection. Second, a multi-source cross-modal penetration fusion network not only accomplishes unidirectional fusion of auxiliary information such as multi-dimensional rationales but also achieves complete mutual complementation between text and the image. Comprehensive experiments demonstrate the effectiveness of M2ERD for rumor detection on Weibo, RumorEval, and Pheme datasets, achieving a 2.6% improvement in accuracy and a 1.9% improvement in F1-score compared to all baselines. We release the code and data at <span><span>https://github.com/youchengyan/M2ERD</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"190 ","pages":"Article 107625"},"PeriodicalIF":6.0000,"publicationDate":"2025-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Collaborate large and small language models for multi-modal emergency rumor detection\",\"authors\":\"Youcheng Yan , Jinshuo Liu , Juan Deng , Junyan Li , Lina Wang , Jeff Z. Pan\",\"doi\":\"10.1016/j.neunet.2025.107625\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Multi-modal emergency rumors are spreading in the current digital era, causing significant disruptions and negative impacts. Most existing methods focus on exploring rumor detection using individual small language models (SLMs) or large language models (LLMs), achieving a certain degree of success but with underlying issues. Approaches based on SLMs have reached a bottleneck due to their limited knowledge and capacity. In contrast, LLMs have unique strengths in deep analysis that compensate for the weaknesses of SLMs; however, they struggle to select and integrate analyses to draw appropriate conclusions. Furthermore, recent works on multi-modal feature fusion remain superficial, limiting the ability of these models to fully comprehend and identify rumors. In this work, we propose Collaborate Large and Small Language Models for Multi-Modal Emergency Rumor Detection (M2ERD). Specifically, it consists of two main components. First, LLMs generate multi-dimensional rationales based on multi-perspective prompts, from which SLMs selectively derive insights for rumor detection. Second, a multi-source cross-modal penetration fusion network not only accomplishes unidirectional fusion of auxiliary information such as multi-dimensional rationales but also achieves complete mutual complementation between text and the image. Comprehensive experiments demonstrate the effectiveness of M2ERD for rumor detection on Weibo, RumorEval, and Pheme datasets, achieving a 2.6% improvement in accuracy and a 1.9% improvement in F1-score compared to all baselines. We release the code and data at <span><span>https://github.com/youchengyan/M2ERD</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":49763,\"journal\":{\"name\":\"Neural Networks\",\"volume\":\"190 \",\"pages\":\"Article 107625\"},\"PeriodicalIF\":6.0000,\"publicationDate\":\"2025-06-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neural Networks\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0893608025005052\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0893608025005052","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Collaborate large and small language models for multi-modal emergency rumor detection
Multi-modal emergency rumors are spreading in the current digital era, causing significant disruptions and negative impacts. Most existing methods focus on exploring rumor detection using individual small language models (SLMs) or large language models (LLMs), achieving a certain degree of success but with underlying issues. Approaches based on SLMs have reached a bottleneck due to their limited knowledge and capacity. In contrast, LLMs have unique strengths in deep analysis that compensate for the weaknesses of SLMs; however, they struggle to select and integrate analyses to draw appropriate conclusions. Furthermore, recent works on multi-modal feature fusion remain superficial, limiting the ability of these models to fully comprehend and identify rumors. In this work, we propose Collaborate Large and Small Language Models for Multi-Modal Emergency Rumor Detection (M2ERD). Specifically, it consists of two main components. First, LLMs generate multi-dimensional rationales based on multi-perspective prompts, from which SLMs selectively derive insights for rumor detection. Second, a multi-source cross-modal penetration fusion network not only accomplishes unidirectional fusion of auxiliary information such as multi-dimensional rationales but also achieves complete mutual complementation between text and the image. Comprehensive experiments demonstrate the effectiveness of M2ERD for rumor detection on Weibo, RumorEval, and Pheme datasets, achieving a 2.6% improvement in accuracy and a 1.9% improvement in F1-score compared to all baselines. We release the code and data at https://github.com/youchengyan/M2ERD.
期刊介绍:
Neural Networks is a platform that aims to foster an international community of scholars and practitioners interested in neural networks, deep learning, and other approaches to artificial intelligence and machine learning. Our journal invites submissions covering various aspects of neural networks research, from computational neuroscience and cognitive modeling to mathematical analyses and engineering applications. By providing a forum for interdisciplinary discussions between biology and technology, we aim to encourage the development of biologically-inspired artificial intelligence.