Enhanced Feature Representation for Multimodal Fake News Detection Using Localized Fine-Tuning of Improved BERT and VGG-19 Models

IF 2.9 4区 综合性期刊 Q1 Multidisciplinary
Suhaib Kh. Hamed, Mohd Juzaiddin Ab Aziz, Mohd Ridzwan Yaakub
{"title":"Enhanced Feature Representation for Multimodal Fake News Detection Using Localized Fine-Tuning of Improved BERT and VGG-19 Models","authors":"Suhaib Kh. Hamed, Mohd Juzaiddin Ab Aziz, Mohd Ridzwan Yaakub","doi":"10.1007/s13369-024-09354-2","DOIUrl":null,"url":null,"abstract":"<p>The spread of fake news poses significant challenges across various sectors, including health, the economy, politics, and national stability. Social media and modern technology have facilitated the rapid dissemination of fake news, predominantly in multimedia formats. Despite advancements, multimodal fake news detection models struggle to achieve optimal accuracy, primarily due to the quality of feature representation. This study aims to enhance feature representation to improve fake news identification. Pre-trained models for feature extraction, typically designed for general public domains, may not suit the specific characteristics of our task using the Fakeddit dataset. We propose a localized fine-tuning strategy, refining pre-trained BERT and VGG-19 models for accurate multimodal feature representation in fake news detection. BERT was fine-tuned by retraining all layers, while only the last block of VGG-19 was fine-tuned. To further enhance the representations, we made structural modifications to VGG-19, including the use of a global average pooling layer and a redesigned classifier. This approach significantly improved our multimodal fake news detection model’s performance, achieving a high accuracy of 92%. Compared to state-of-the-art studies that use generic pre-trained models, our model demonstrates superior performance. Our research underscores the importance of feature representation in multimodal contexts and opens avenues for exploring the synergy between textual and visual modalities in fake news detection.</p>","PeriodicalId":8109,"journal":{"name":"Arabian Journal for Science and Engineering","volume":null,"pages":null},"PeriodicalIF":2.9000,"publicationDate":"2024-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Arabian Journal for Science and Engineering","FirstCategoryId":"103","ListUrlMain":"https://doi.org/10.1007/s13369-024-09354-2","RegionNum":4,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"Multidisciplinary","Score":null,"Total":0}
引用次数: 0

Abstract

The spread of fake news poses significant challenges across various sectors, including health, the economy, politics, and national stability. Social media and modern technology have facilitated the rapid dissemination of fake news, predominantly in multimedia formats. Despite advancements, multimodal fake news detection models struggle to achieve optimal accuracy, primarily due to the quality of feature representation. This study aims to enhance feature representation to improve fake news identification. Pre-trained models for feature extraction, typically designed for general public domains, may not suit the specific characteristics of our task using the Fakeddit dataset. We propose a localized fine-tuning strategy, refining pre-trained BERT and VGG-19 models for accurate multimodal feature representation in fake news detection. BERT was fine-tuned by retraining all layers, while only the last block of VGG-19 was fine-tuned. To further enhance the representations, we made structural modifications to VGG-19, including the use of a global average pooling layer and a redesigned classifier. This approach significantly improved our multimodal fake news detection model’s performance, achieving a high accuracy of 92%. Compared to state-of-the-art studies that use generic pre-trained models, our model demonstrates superior performance. Our research underscores the importance of feature representation in multimodal contexts and opens avenues for exploring the synergy between textual and visual modalities in fake news detection.

Abstract Image

利用改进型 BERT 和 VGG-19 模型的局部微调增强多模态假新闻检测的特征表示
假新闻的传播给卫生、经济、政治和国家稳定等各个领域带来了重大挑战。社交媒体和现代技术促进了以多媒体形式为主的假新闻的快速传播。尽管取得了进步,多模态假新闻检测模型仍难以达到最佳准确度,这主要是由于特征表示的质量造成的。本研究旨在增强特征表示,以提高假新闻识别率。用于特征提取的预训练模型通常是为一般公共领域设计的,可能不适合我们使用 Fakeddit 数据集的任务的具体特点。我们提出了一种局部微调策略,对预先训练的 BERT 和 VGG-19 模型进行改进,从而在假新闻检测中实现准确的多模态特征表示。通过重新训练所有层对 BERT 进行了微调,而只对 VGG-19 的最后一个区块进行了微调。为了进一步增强表征,我们对 VGG-19 进行了结构性修改,包括使用全局平均池化层和重新设计的分类器。这种方法大大提高了多模态假新闻检测模型的性能,准确率高达 92%。与使用通用预训练模型的最先进研究相比,我们的模型表现出了卓越的性能。我们的研究强调了多模态背景下特征表示的重要性,并为探索假新闻检测中文本和视觉模态之间的协同作用开辟了途径。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Arabian Journal for Science and Engineering
Arabian Journal for Science and Engineering 综合性期刊-综合性期刊
CiteScore
5.20
自引率
3.40%
发文量
0
审稿时长
4.3 months
期刊介绍: King Fahd University of Petroleum & Minerals (KFUPM) partnered with Springer to publish the Arabian Journal for Science and Engineering (AJSE). AJSE, which has been published by KFUPM since 1975, is a recognized national, regional and international journal that provides a great opportunity for the dissemination of research advances from the Kingdom of Saudi Arabia, MENA and the world.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信