Ankit Kumar , Mikail Mohammed Salim , David Camacho , Jong Hyuk Park
{"title":"多媒体数据安全的大型语言模型:挑战与解决方案综述","authors":"Ankit Kumar , Mikail Mohammed Salim , David Camacho , Jong Hyuk Park","doi":"10.1016/j.comnet.2025.111379","DOIUrl":null,"url":null,"abstract":"<div><div>The rapid expansion of IoT applications utilizes multimedia data integrated with Large Language Models (LLMs) for interpreting digital information by leveraging the capabilities of artificial intelligence (AI) driven neural network systems. These models are extensively used as generative AI tools for data augmentation but data security and privacy remain a fundamental concern associated with LLM model in the digital domain. Traditional security approach shows potential challenges in addressing emerging threats such as adversarial attacks, data poisoning, or privacy breaches, especially in dynamic and resource-constrained IoT environments. Such malicious attacks target the LLM model during the learning and evaluation phase to exploit the vulnerabilities for unauthorized access. The proposed study conducts a comprehensive survey of the transformative potential of LLM models for securing multimedia data offering analysis of their capabilities, challenges, and solutions. The proposed study explores potential security threats and remedies for each type of multimedia data and investigates the various traditional and emerging data protection schemes. The study systematically classifies emerging attacks on LLM models during training and testing phases which include membership attacks, adversarial perturbations, prompt injection, etc. The study also investigates the various robust defense mechanism such as adversarial training, regularization, encryption, etc. The study evaluates the efficiency of potential LLM models such as generative LLM, transformer-based, and other multimodal systems in securing image, text, and video multimedia data highlighting their adaptability and scalability. The proposed survey compares state-of-the-art solutions and underscores the efficiency of LLM-driven mechanisms over traditional approaches in mitigating emerging attacks such as zero-day threats on multimedia data. It ensures real-time compliance with standard regulations like GDPR (General Data Protection Regulation). The proposed work identifies some open challenges including privacy-preserving LLM deployment, black-box interpretability, personalized LLM privacy risk, and cross-model security integration. It also highlights some robust future solutions such as lightweight LLM design and hybrid security frameworks. The proposed work bridges critical research gaps by providing insights into LLM-based emerging techniques to safeguard sensitive data in IoT-based real-world applications.</div></div>","PeriodicalId":50637,"journal":{"name":"Computer Networks","volume":"267 ","pages":"Article 111379"},"PeriodicalIF":4.4000,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A comprehensive survey on large language models for multimedia data security: challenges and solutions\",\"authors\":\"Ankit Kumar , Mikail Mohammed Salim , David Camacho , Jong Hyuk Park\",\"doi\":\"10.1016/j.comnet.2025.111379\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The rapid expansion of IoT applications utilizes multimedia data integrated with Large Language Models (LLMs) for interpreting digital information by leveraging the capabilities of artificial intelligence (AI) driven neural network systems. These models are extensively used as generative AI tools for data augmentation but data security and privacy remain a fundamental concern associated with LLM model in the digital domain. Traditional security approach shows potential challenges in addressing emerging threats such as adversarial attacks, data poisoning, or privacy breaches, especially in dynamic and resource-constrained IoT environments. Such malicious attacks target the LLM model during the learning and evaluation phase to exploit the vulnerabilities for unauthorized access. The proposed study conducts a comprehensive survey of the transformative potential of LLM models for securing multimedia data offering analysis of their capabilities, challenges, and solutions. The proposed study explores potential security threats and remedies for each type of multimedia data and investigates the various traditional and emerging data protection schemes. The study systematically classifies emerging attacks on LLM models during training and testing phases which include membership attacks, adversarial perturbations, prompt injection, etc. The study also investigates the various robust defense mechanism such as adversarial training, regularization, encryption, etc. The study evaluates the efficiency of potential LLM models such as generative LLM, transformer-based, and other multimodal systems in securing image, text, and video multimedia data highlighting their adaptability and scalability. The proposed survey compares state-of-the-art solutions and underscores the efficiency of LLM-driven mechanisms over traditional approaches in mitigating emerging attacks such as zero-day threats on multimedia data. It ensures real-time compliance with standard regulations like GDPR (General Data Protection Regulation). The proposed work identifies some open challenges including privacy-preserving LLM deployment, black-box interpretability, personalized LLM privacy risk, and cross-model security integration. It also highlights some robust future solutions such as lightweight LLM design and hybrid security frameworks. The proposed work bridges critical research gaps by providing insights into LLM-based emerging techniques to safeguard sensitive data in IoT-based real-world applications.</div></div>\",\"PeriodicalId\":50637,\"journal\":{\"name\":\"Computer Networks\",\"volume\":\"267 \",\"pages\":\"Article 111379\"},\"PeriodicalIF\":4.4000,\"publicationDate\":\"2025-05-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computer Networks\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1389128625003469\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1389128625003469","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
A comprehensive survey on large language models for multimedia data security: challenges and solutions
The rapid expansion of IoT applications utilizes multimedia data integrated with Large Language Models (LLMs) for interpreting digital information by leveraging the capabilities of artificial intelligence (AI) driven neural network systems. These models are extensively used as generative AI tools for data augmentation but data security and privacy remain a fundamental concern associated with LLM model in the digital domain. Traditional security approach shows potential challenges in addressing emerging threats such as adversarial attacks, data poisoning, or privacy breaches, especially in dynamic and resource-constrained IoT environments. Such malicious attacks target the LLM model during the learning and evaluation phase to exploit the vulnerabilities for unauthorized access. The proposed study conducts a comprehensive survey of the transformative potential of LLM models for securing multimedia data offering analysis of their capabilities, challenges, and solutions. The proposed study explores potential security threats and remedies for each type of multimedia data and investigates the various traditional and emerging data protection schemes. The study systematically classifies emerging attacks on LLM models during training and testing phases which include membership attacks, adversarial perturbations, prompt injection, etc. The study also investigates the various robust defense mechanism such as adversarial training, regularization, encryption, etc. The study evaluates the efficiency of potential LLM models such as generative LLM, transformer-based, and other multimodal systems in securing image, text, and video multimedia data highlighting their adaptability and scalability. The proposed survey compares state-of-the-art solutions and underscores the efficiency of LLM-driven mechanisms over traditional approaches in mitigating emerging attacks such as zero-day threats on multimedia data. It ensures real-time compliance with standard regulations like GDPR (General Data Protection Regulation). The proposed work identifies some open challenges including privacy-preserving LLM deployment, black-box interpretability, personalized LLM privacy risk, and cross-model security integration. It also highlights some robust future solutions such as lightweight LLM design and hybrid security frameworks. The proposed work bridges critical research gaps by providing insights into LLM-based emerging techniques to safeguard sensitive data in IoT-based real-world applications.
期刊介绍:
Computer Networks is an international, archival journal providing a publication vehicle for complete coverage of all topics of interest to those involved in the computer communications networking area. The audience includes researchers, managers and operators of networks as well as designers and implementors. The Editorial Board will consider any material for publication that is of interest to those groups.