Fused Swish-ReLU Efficient-Net Model for Deepfakes Detection

Hafsa Ilyas, A. Javed, Muteb Aljasem, Mustafa Alhababi
{"title":"Fused Swish-ReLU Efficient-Net Model for Deepfakes Detection","authors":"Hafsa Ilyas, A. Javed, Muteb Aljasem, Mustafa Alhababi","doi":"10.1109/ICARA56516.2023.10125801","DOIUrl":null,"url":null,"abstract":"With the rapid development of sophisticated deepfakes generation methods, the realism of fake content has reached the level where it becomes difficult for human eyes to identify such high-quality fake images/videos, thus increasing the demand for developing deepfakes detection methods. The diversity in deepfakes images/videos in terms of ethnicity, illumination condition, skin tone, age, background setting, and generation algorithms makes the detection task quite difficult. To better address the aforementioned challenges, we present a novel Swish-ReLU Efficient-Net (SRE-Net) that is robust to the identification of deepfakes generated using different face-swap and face-reenactment techniques. More precisely, we fused two EfficienNet-b0 models, one with the ReLU and the other with the Swish activation function along with layer freezing to achieve better detection results. Our SRE-Net attained the average accuracy and precision of 96.5% and 97.07% on the FaceForensics++ dataset, and 88.41% and 91.28% on the DFDC-preview dataset. The high detection results demonstrate the effectiveness of SRE-Net while detecting the deepfakes generated using different manipulation algorithms.","PeriodicalId":443572,"journal":{"name":"2023 9th International Conference on Automation, Robotics and Applications (ICARA)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 9th International Conference on Automation, Robotics and Applications (ICARA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICARA56516.2023.10125801","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

With the rapid development of sophisticated deepfakes generation methods, the realism of fake content has reached the level where it becomes difficult for human eyes to identify such high-quality fake images/videos, thus increasing the demand for developing deepfakes detection methods. The diversity in deepfakes images/videos in terms of ethnicity, illumination condition, skin tone, age, background setting, and generation algorithms makes the detection task quite difficult. To better address the aforementioned challenges, we present a novel Swish-ReLU Efficient-Net (SRE-Net) that is robust to the identification of deepfakes generated using different face-swap and face-reenactment techniques. More precisely, we fused two EfficienNet-b0 models, one with the ReLU and the other with the Swish activation function along with layer freezing to achieve better detection results. Our SRE-Net attained the average accuracy and precision of 96.5% and 97.07% on the FaceForensics++ dataset, and 88.41% and 91.28% on the DFDC-preview dataset. The high detection results demonstrate the effectiveness of SRE-Net while detecting the deepfakes generated using different manipulation algorithms.
融合Swish-ReLU高效网络模型的深度伪造检测
随着复杂的深度伪造生成方法的快速发展,虚假内容的真实感已经达到人眼难以识别高质量的假图像/视频的程度,从而增加了开发深度伪造检测方法的需求。深度伪造图像/视频在种族、光照条件、肤色、年龄、背景设置和生成算法方面的多样性使得检测任务相当困难。为了更好地解决上述挑战,我们提出了一种新的Swish-ReLU Efficient-Net (SRE-Net),它对使用不同的人脸交换和人脸再现技术生成的深度伪造具有鲁棒性。更准确地说,我们融合了两个EfficienNet-b0模型,一个带有ReLU,另一个带有Swish激活功能以及层冻结,以获得更好的检测结果。我们的SRE-Net在FaceForensics++数据集上的平均准确率和精密度分别为96.5%和97.07%,在DFDC-preview数据集上的平均准确率和精密度分别为88.41%和91.28%。高检测结果证明了SRE-Net在检测使用不同操作算法生成的深度伪造时的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信