{"title":"Framework for Quantum-Based Deepfake Video Detection (Without Audio)","authors":"Atul Pandey, Bhawana Rudra, Rajesh Kumar Krishnan","doi":"10.1155/int/3990069","DOIUrl":null,"url":null,"abstract":"<div>\n <p>Artificial intelligence (AI) has made human tasks easier compared to earlier days. It has revolutionized various domains, from paper drafting to video editing. However, some individuals exploit AI to create deceptive content, such as fake videos, audios, and images, to mislead others. To address this, researchers and large corporations have proposed solutions for detecting fake content using classical deep learning models. However, these models often suffer from a large number of trainable parameters, which leads to large model sizes and, consequently, computational intensive. To overcome these limitations, we propose various hybrid classical–quantum models that use a classical pre-trained model as a front-end feature extractor, followed by a quantum-based LSTM network, that is, QLSTM. These pre-trained models are based on the ResNet architecture, such as ResNet34, 50, and 101. We have compared the performance of the proposed models with their classical counterparts. These proposed models combine the strengths of classical and quantum systems for the detection of deepfake video (without audio). Our results indicate that the proposed models significantly reduce the number of trainable parameters, as well as quantum long short-term memory (QLSTM) parameters, which leads to a smaller model size than the classical models. Despite the reduced parameter, the performance of the proposed models is either superior to or comparable with that of their classical equivalent. The proposed hybrid quantum models, that is, ResNet34-QLSTM, ResNet50-QLSTM, and ResNet101-QLSTM, achieve a reduction of approximately 1.50%, 4.59%, and 5.24% in total trainable parameters compared to their equivalent classical models, respectively. Additionally, QLSTM linked with the proposed models reduces its trainable parameters by 99.02%, 99.16%, and 99.55%, respectively, compared to equivalent classical LSTM. This significant reduction highlights the efficiency of the quantum-based network in terms of resource usage. The trained model sizes of the proposed models are 81.35, 88.06, and 162.79, and their equivalent classical models are 82.59, 92.28, and 171.76 in MB, respectively.</p>\n </div>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":"2025 1","pages":""},"PeriodicalIF":5.0000,"publicationDate":"2025-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1155/int/3990069","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Intelligent Systems","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1155/int/3990069","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Artificial intelligence (AI) has made human tasks easier compared to earlier days. It has revolutionized various domains, from paper drafting to video editing. However, some individuals exploit AI to create deceptive content, such as fake videos, audios, and images, to mislead others. To address this, researchers and large corporations have proposed solutions for detecting fake content using classical deep learning models. However, these models often suffer from a large number of trainable parameters, which leads to large model sizes and, consequently, computational intensive. To overcome these limitations, we propose various hybrid classical–quantum models that use a classical pre-trained model as a front-end feature extractor, followed by a quantum-based LSTM network, that is, QLSTM. These pre-trained models are based on the ResNet architecture, such as ResNet34, 50, and 101. We have compared the performance of the proposed models with their classical counterparts. These proposed models combine the strengths of classical and quantum systems for the detection of deepfake video (without audio). Our results indicate that the proposed models significantly reduce the number of trainable parameters, as well as quantum long short-term memory (QLSTM) parameters, which leads to a smaller model size than the classical models. Despite the reduced parameter, the performance of the proposed models is either superior to or comparable with that of their classical equivalent. The proposed hybrid quantum models, that is, ResNet34-QLSTM, ResNet50-QLSTM, and ResNet101-QLSTM, achieve a reduction of approximately 1.50%, 4.59%, and 5.24% in total trainable parameters compared to their equivalent classical models, respectively. Additionally, QLSTM linked with the proposed models reduces its trainable parameters by 99.02%, 99.16%, and 99.55%, respectively, compared to equivalent classical LSTM. This significant reduction highlights the efficiency of the quantum-based network in terms of resource usage. The trained model sizes of the proposed models are 81.35, 88.06, and 162.79, and their equivalent classical models are 82.59, 92.28, and 171.76 in MB, respectively.
期刊介绍:
The International Journal of Intelligent Systems serves as a forum for individuals interested in tapping into the vast theories based on intelligent systems construction. With its peer-reviewed format, the journal explores several fascinating editorials written by today''s experts in the field. Because new developments are being introduced each day, there''s much to be learned — examination, analysis creation, information retrieval, man–computer interactions, and more. The International Journal of Intelligent Systems uses charts and illustrations to demonstrate these ground-breaking issues, and encourages readers to share their thoughts and experiences.