{"title":"开发先进的机器学习模型,用于识别数字媒体中的虚拟蒙太奇","authors":"Yuxuan Liu","doi":"10.1016/j.entcom.2025.101002","DOIUrl":null,"url":null,"abstract":"<div><div>Virtual montages are compositions that blend several digital materials to produce a new, visually appealing narrative and are becoming increasingly popular, as a result of the quick development of digital media. Virtual montages have emerged as a potent technique for improving visual storytelling and content presentation across platforms in the digital age. However, the quick expansion of digital content has made advanced methods for effectively recognizing and classifying these montages necessary. This work suggests a novel method for virtual montage identification across a broad range of visual themes, content types, and media resolutions, and resolutions utilizing an Adaptive Flower Pollination Optimized Mutual Information with Naïve Bayes (AFPO-MI-NB) model. The research’s dataset comprises a variety of content from virtual montage detection dataset which enables the model to generalize effectively to data from the real world. To improve the model’s capacity to handle various image and video qualities, pre-processing methods, including data augmentation and pixel value normalization are used. Convolutional neural networks (CNN) are used for feature extraction to capture spatial patterns. The model uses AFPO-MI-NB classification to enhance classification accuracy and computational efficiency. AFPO allows for more adaptable feature weights, MI evaluates the relationship between visual features and the classification task, and NB classifier processes feature matrices for binary classification. The hybrid approach strengthens feature selection through global and local searches, resulting in a model that enhances classification accuracy and improves computational efficiency. According to the experimental data, this model provides a reliable solution for virtual montage identification across a variety of media formats, outperforming current techniques in terms of F1-score of 0.95, recall of 0.92, and precision of 0.97. This work has significant applications in automated digital media analysis, copyright enforcement, and content control.</div></div>","PeriodicalId":55997,"journal":{"name":"Entertainment Computing","volume":"55 ","pages":"Article 101002"},"PeriodicalIF":2.4000,"publicationDate":"2025-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Developing an advanced machine learning model for identifying virtual montages in digital media\",\"authors\":\"Yuxuan Liu\",\"doi\":\"10.1016/j.entcom.2025.101002\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Virtual montages are compositions that blend several digital materials to produce a new, visually appealing narrative and are becoming increasingly popular, as a result of the quick development of digital media. Virtual montages have emerged as a potent technique for improving visual storytelling and content presentation across platforms in the digital age. However, the quick expansion of digital content has made advanced methods for effectively recognizing and classifying these montages necessary. This work suggests a novel method for virtual montage identification across a broad range of visual themes, content types, and media resolutions, and resolutions utilizing an Adaptive Flower Pollination Optimized Mutual Information with Naïve Bayes (AFPO-MI-NB) model. The research’s dataset comprises a variety of content from virtual montage detection dataset which enables the model to generalize effectively to data from the real world. To improve the model’s capacity to handle various image and video qualities, pre-processing methods, including data augmentation and pixel value normalization are used. Convolutional neural networks (CNN) are used for feature extraction to capture spatial patterns. The model uses AFPO-MI-NB classification to enhance classification accuracy and computational efficiency. AFPO allows for more adaptable feature weights, MI evaluates the relationship between visual features and the classification task, and NB classifier processes feature matrices for binary classification. The hybrid approach strengthens feature selection through global and local searches, resulting in a model that enhances classification accuracy and improves computational efficiency. According to the experimental data, this model provides a reliable solution for virtual montage identification across a variety of media formats, outperforming current techniques in terms of F1-score of 0.95, recall of 0.92, and precision of 0.97. This work has significant applications in automated digital media analysis, copyright enforcement, and content control.</div></div>\",\"PeriodicalId\":55997,\"journal\":{\"name\":\"Entertainment Computing\",\"volume\":\"55 \",\"pages\":\"Article 101002\"},\"PeriodicalIF\":2.4000,\"publicationDate\":\"2025-07-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Entertainment Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1875952125000825\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, CYBERNETICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Entertainment Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1875952125000825","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, CYBERNETICS","Score":null,"Total":0}
Developing an advanced machine learning model for identifying virtual montages in digital media
Virtual montages are compositions that blend several digital materials to produce a new, visually appealing narrative and are becoming increasingly popular, as a result of the quick development of digital media. Virtual montages have emerged as a potent technique for improving visual storytelling and content presentation across platforms in the digital age. However, the quick expansion of digital content has made advanced methods for effectively recognizing and classifying these montages necessary. This work suggests a novel method for virtual montage identification across a broad range of visual themes, content types, and media resolutions, and resolutions utilizing an Adaptive Flower Pollination Optimized Mutual Information with Naïve Bayes (AFPO-MI-NB) model. The research’s dataset comprises a variety of content from virtual montage detection dataset which enables the model to generalize effectively to data from the real world. To improve the model’s capacity to handle various image and video qualities, pre-processing methods, including data augmentation and pixel value normalization are used. Convolutional neural networks (CNN) are used for feature extraction to capture spatial patterns. The model uses AFPO-MI-NB classification to enhance classification accuracy and computational efficiency. AFPO allows for more adaptable feature weights, MI evaluates the relationship between visual features and the classification task, and NB classifier processes feature matrices for binary classification. The hybrid approach strengthens feature selection through global and local searches, resulting in a model that enhances classification accuracy and improves computational efficiency. According to the experimental data, this model provides a reliable solution for virtual montage identification across a variety of media formats, outperforming current techniques in terms of F1-score of 0.95, recall of 0.92, and precision of 0.97. This work has significant applications in automated digital media analysis, copyright enforcement, and content control.
期刊介绍:
Entertainment Computing publishes original, peer-reviewed research articles and serves as a forum for stimulating and disseminating innovative research ideas, emerging technologies, empirical investigations, state-of-the-art methods and tools in all aspects of digital entertainment, new media, entertainment computing, gaming, robotics, toys and applications among researchers, engineers, social scientists, artists and practitioners. Theoretical, technical, empirical, survey articles and case studies are all appropriate to the journal.