Pir Noman Ahmad , Adnan Muhammad Shah , KangYoon Lee , Wazir Muhammad
{"title":"基于预训练语言模型的在线社交网络错误信息检测","authors":"Pir Noman Ahmad , Adnan Muhammad Shah , KangYoon Lee , Wazir Muhammad","doi":"10.1016/j.ipm.2025.104342","DOIUrl":null,"url":null,"abstract":"<div><div>The growing prevalence of online misinformation poses substantial threats, with notable examples including the undermined integrity of democratic processes and decreased effectiveness of public health efforts. The effectiveness of existing solutions, such as user education and content removal, remains unclear, primarily because confirmation bias and peer pressure hinder the identification of noncredible information by users. To address these challenges posed by online misinformation, this study proposes a state-of-the-art approach that leverages transformer-based models, including bidirectional encoder representation from transformers (BERT), GPT-2, and XLNet. These models leverage attention mechanisms to simultaneously process and capture contextual subtleties in documents, enabling highly accurate misinformation detection and classification in dynamic and complex online narratives. A transformer-based pretrained language model is used to analyze, a large corpus of tweets related to misinformation events concerning the 2020 U.S. election. Although isolated interventions are found to be ineffective, a synergistic approach is shown to reduce misinformation prevalence by 87.9 % within a 40-min delay based on a credibility interval of 80 %. These findings highlight the potential of empirical models to inform policies, enhance content moderation practices, and strengthen public resilience against misinformation.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 1","pages":"Article 104342"},"PeriodicalIF":6.9000,"publicationDate":"2025-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Misinformation detection on online social networks using pretrained language models\",\"authors\":\"Pir Noman Ahmad , Adnan Muhammad Shah , KangYoon Lee , Wazir Muhammad\",\"doi\":\"10.1016/j.ipm.2025.104342\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The growing prevalence of online misinformation poses substantial threats, with notable examples including the undermined integrity of democratic processes and decreased effectiveness of public health efforts. The effectiveness of existing solutions, such as user education and content removal, remains unclear, primarily because confirmation bias and peer pressure hinder the identification of noncredible information by users. To address these challenges posed by online misinformation, this study proposes a state-of-the-art approach that leverages transformer-based models, including bidirectional encoder representation from transformers (BERT), GPT-2, and XLNet. These models leverage attention mechanisms to simultaneously process and capture contextual subtleties in documents, enabling highly accurate misinformation detection and classification in dynamic and complex online narratives. A transformer-based pretrained language model is used to analyze, a large corpus of tweets related to misinformation events concerning the 2020 U.S. election. Although isolated interventions are found to be ineffective, a synergistic approach is shown to reduce misinformation prevalence by 87.9 % within a 40-min delay based on a credibility interval of 80 %. These findings highlight the potential of empirical models to inform policies, enhance content moderation practices, and strengthen public resilience against misinformation.</div></div>\",\"PeriodicalId\":50365,\"journal\":{\"name\":\"Information Processing & Management\",\"volume\":\"63 1\",\"pages\":\"Article 104342\"},\"PeriodicalIF\":6.9000,\"publicationDate\":\"2025-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Processing & Management\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0306457325002833\",\"RegionNum\":1,\"RegionCategory\":\"管理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Processing & Management","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306457325002833","RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Misinformation detection on online social networks using pretrained language models
The growing prevalence of online misinformation poses substantial threats, with notable examples including the undermined integrity of democratic processes and decreased effectiveness of public health efforts. The effectiveness of existing solutions, such as user education and content removal, remains unclear, primarily because confirmation bias and peer pressure hinder the identification of noncredible information by users. To address these challenges posed by online misinformation, this study proposes a state-of-the-art approach that leverages transformer-based models, including bidirectional encoder representation from transformers (BERT), GPT-2, and XLNet. These models leverage attention mechanisms to simultaneously process and capture contextual subtleties in documents, enabling highly accurate misinformation detection and classification in dynamic and complex online narratives. A transformer-based pretrained language model is used to analyze, a large corpus of tweets related to misinformation events concerning the 2020 U.S. election. Although isolated interventions are found to be ineffective, a synergistic approach is shown to reduce misinformation prevalence by 87.9 % within a 40-min delay based on a credibility interval of 80 %. These findings highlight the potential of empirical models to inform policies, enhance content moderation practices, and strengthen public resilience against misinformation.
期刊介绍:
Information Processing and Management is dedicated to publishing cutting-edge original research at the convergence of computing and information science. Our scope encompasses theory, methods, and applications across various domains, including advertising, business, health, information science, information technology marketing, and social computing.
We aim to cater to the interests of both primary researchers and practitioners by offering an effective platform for the timely dissemination of advanced and topical issues in this interdisciplinary field. The journal places particular emphasis on original research articles, research survey articles, research method articles, and articles addressing critical applications of research. Join us in advancing knowledge and innovation at the intersection of computing and information science.