{"title":"A novel method for real-time object-based copy-move tampering localization in videos using fine-tuned YOLO V8","authors":"Sandhya, Abhishek Kashyap","doi":"10.1016/j.fsidi.2023.301663","DOIUrl":null,"url":null,"abstract":"<div><p>The research community faces challenges for video forgery detection techniques as advancements in multimedia technology have made it easy to alter the original video content and share it on electronic and social media with false propaganda. The copy-move attack is the most commonly practiced type of attack in videos/images, where an object is copied and moved into the current frame or any other frame of the video. Hence an illusion of recreation can be created to forge the content. It is very difficult to differentiate to uncover the forgery traces by the naked eye. Hence, a passive method-based algorithm is proposed to scientifically investigate the statistical properties of the video by normalizing the median difference of the frames at the pixel level, and graphical analysis successfully shows the clear peak in the forged region. After that, a new deep learning approach, “You Only Look at Once”, the latest eighth version of YOLO, is tuned and trained for the localization of forged objects in the real-time domain. The validation and testing results obtained from the trained YOLO V8 are successfully able to detect and localize the forged objects in the videos with mean average precision (mAP) of 0.99, recall is 0.99, precision is 0.99, and highest confidence score. The proposed YOLO V8 is fine-tuned in three different ways, and the performance of the proposed method outperforms existing state-of-the-art techniques in terms of inference speed, accuracy, precision, recall, testing, and training time.</p></div>","PeriodicalId":48481,"journal":{"name":"Forensic Science International-Digital Investigation","volume":"48 ","pages":"Article 301663"},"PeriodicalIF":2.0000,"publicationDate":"2023-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666281723001828/pdfft?md5=35ac6006d6528037ce8427b12e149b58&pid=1-s2.0-S2666281723001828-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Forensic Science International-Digital Investigation","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666281723001828","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
The research community faces challenges for video forgery detection techniques as advancements in multimedia technology have made it easy to alter the original video content and share it on electronic and social media with false propaganda. The copy-move attack is the most commonly practiced type of attack in videos/images, where an object is copied and moved into the current frame or any other frame of the video. Hence an illusion of recreation can be created to forge the content. It is very difficult to differentiate to uncover the forgery traces by the naked eye. Hence, a passive method-based algorithm is proposed to scientifically investigate the statistical properties of the video by normalizing the median difference of the frames at the pixel level, and graphical analysis successfully shows the clear peak in the forged region. After that, a new deep learning approach, “You Only Look at Once”, the latest eighth version of YOLO, is tuned and trained for the localization of forged objects in the real-time domain. The validation and testing results obtained from the trained YOLO V8 are successfully able to detect and localize the forged objects in the videos with mean average precision (mAP) of 0.99, recall is 0.99, precision is 0.99, and highest confidence score. The proposed YOLO V8 is fine-tuned in three different ways, and the performance of the proposed method outperforms existing state-of-the-art techniques in terms of inference speed, accuracy, precision, recall, testing, and training time.