Jinglong Duan , Shiqing Wu , Weihua Li , Quan Bai , Minh Nguyen , Jianhua Jiang
{"title":"BotDMM: Dual-channel multi-modal learning for LLM-driven bot detection on social media","authors":"Jinglong Duan , Shiqing Wu , Weihua Li , Quan Bai , Minh Nguyen , Jianhua Jiang","doi":"10.1016/j.inffus.2025.103758","DOIUrl":null,"url":null,"abstract":"<div><div>Social bots are becoming a growing concern due to their ability to spread misinformation and manipulate public discourse. The emergence of powerful Large Language Models (LLMs), such as ChatGPT, has introduced a new generation of bots capable of producing fluent and human-like text while dynamically adapting their relational patterns over time. These LLM-driven bots seamlessly blend into online communities, making them significantly more challenging to detect. Most existing approaches rely on static features or simple behavioral patterns, which are not effective against bots that can evolve both their language and their network connections. To address these challenges, we propose a novel Dual-channel Multi-Modal learning (BotDMM) framework for LLM-driven bot detection. The proposed model captures discriminative information from two complementary sources: users’ content features (including their profiles and temporal posting behavior) and structural features (reflecting local network topology). Furthermore, we employ a joint training approach that incorporates two carefully designed self-supervised learning paradigms alongside the primary prediction task to enhance discrimination between human users, traditional bots, and LLM-driven bots. Extensive experiments demonstrate the effectiveness and superiority of BotDMM compared to state-of-the-art baselines. The implementation of BotDMM has been released at: <span><span>https://github.com/JaydenDuan/DualChannelBotDMM</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"127 ","pages":"Article 103758"},"PeriodicalIF":15.5000,"publicationDate":"2025-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525008206","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Social bots are becoming a growing concern due to their ability to spread misinformation and manipulate public discourse. The emergence of powerful Large Language Models (LLMs), such as ChatGPT, has introduced a new generation of bots capable of producing fluent and human-like text while dynamically adapting their relational patterns over time. These LLM-driven bots seamlessly blend into online communities, making them significantly more challenging to detect. Most existing approaches rely on static features or simple behavioral patterns, which are not effective against bots that can evolve both their language and their network connections. To address these challenges, we propose a novel Dual-channel Multi-Modal learning (BotDMM) framework for LLM-driven bot detection. The proposed model captures discriminative information from two complementary sources: users’ content features (including their profiles and temporal posting behavior) and structural features (reflecting local network topology). Furthermore, we employ a joint training approach that incorporates two carefully designed self-supervised learning paradigms alongside the primary prediction task to enhance discrimination between human users, traditional bots, and LLM-driven bots. Extensive experiments demonstrate the effectiveness and superiority of BotDMM compared to state-of-the-art baselines. The implementation of BotDMM has been released at: https://github.com/JaydenDuan/DualChannelBotDMM.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.