{"title":"使用可解释的深度学习实现强大的android恶意软件检测模型","authors":"Masumeh Najibi, Amir Jalaly Bidgoly","doi":"10.1016/j.jisa.2025.104191","DOIUrl":null,"url":null,"abstract":"<div><div>The growing threat of Android malware demands effective and trustworthy detection mechanisms. This paper investigates the robustness of explainable deep learning models for Android malware detection and classification using network flow features. Three deep learning architectures — DNN, 1D-CNN, and BiLSTM — were evaluated on the CICAndMal2017 dataset, with BiLSTM achieving the best performance on unseen samples. Model decisions were analyzed using LIME and SHAP to identify influential and potentially manipulable features. Using domain knowledge, features were categorized based on their resistance to evasion, with emphasis on robust indicators such as TCP flags and initial window sizes. Retraining models using only these robust features resulted in minimal performance degradation while significantly improving explainability and resilience to evasion. On the unseen dataset, the BiLSTM model achieved a 70.90% F1-score for malware detection and 62.84% for classification, with AUC scores of 73.39% and 79.96%, respectively. After removing weak features, the retrained detection model maintained a 71% F1-score, and the classification model achieved 57%, demonstrating that robustness can be improved without major loss in performance. These results highlight the potential for transparent and dependable AI-driven cybersecurity solutions, particularly in adversarial settings where evasion is common. By emphasizing explainability and robustness, this work contributes towards models that balance performance with trust in evolving threat landscapes.</div></div>","PeriodicalId":48638,"journal":{"name":"Journal of Information Security and Applications","volume":"93 ","pages":"Article 104191"},"PeriodicalIF":3.7000,"publicationDate":"2025-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Towards a robust android malware detection model using explainable deep learning\",\"authors\":\"Masumeh Najibi, Amir Jalaly Bidgoly\",\"doi\":\"10.1016/j.jisa.2025.104191\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The growing threat of Android malware demands effective and trustworthy detection mechanisms. This paper investigates the robustness of explainable deep learning models for Android malware detection and classification using network flow features. Three deep learning architectures — DNN, 1D-CNN, and BiLSTM — were evaluated on the CICAndMal2017 dataset, with BiLSTM achieving the best performance on unseen samples. Model decisions were analyzed using LIME and SHAP to identify influential and potentially manipulable features. Using domain knowledge, features were categorized based on their resistance to evasion, with emphasis on robust indicators such as TCP flags and initial window sizes. Retraining models using only these robust features resulted in minimal performance degradation while significantly improving explainability and resilience to evasion. On the unseen dataset, the BiLSTM model achieved a 70.90% F1-score for malware detection and 62.84% for classification, with AUC scores of 73.39% and 79.96%, respectively. After removing weak features, the retrained detection model maintained a 71% F1-score, and the classification model achieved 57%, demonstrating that robustness can be improved without major loss in performance. These results highlight the potential for transparent and dependable AI-driven cybersecurity solutions, particularly in adversarial settings where evasion is common. By emphasizing explainability and robustness, this work contributes towards models that balance performance with trust in evolving threat landscapes.</div></div>\",\"PeriodicalId\":48638,\"journal\":{\"name\":\"Journal of Information Security and Applications\",\"volume\":\"93 \",\"pages\":\"Article 104191\"},\"PeriodicalIF\":3.7000,\"publicationDate\":\"2025-08-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Information Security and Applications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2214212625002285\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Information Security and Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2214212625002285","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Towards a robust android malware detection model using explainable deep learning
The growing threat of Android malware demands effective and trustworthy detection mechanisms. This paper investigates the robustness of explainable deep learning models for Android malware detection and classification using network flow features. Three deep learning architectures — DNN, 1D-CNN, and BiLSTM — were evaluated on the CICAndMal2017 dataset, with BiLSTM achieving the best performance on unseen samples. Model decisions were analyzed using LIME and SHAP to identify influential and potentially manipulable features. Using domain knowledge, features were categorized based on their resistance to evasion, with emphasis on robust indicators such as TCP flags and initial window sizes. Retraining models using only these robust features resulted in minimal performance degradation while significantly improving explainability and resilience to evasion. On the unseen dataset, the BiLSTM model achieved a 70.90% F1-score for malware detection and 62.84% for classification, with AUC scores of 73.39% and 79.96%, respectively. After removing weak features, the retrained detection model maintained a 71% F1-score, and the classification model achieved 57%, demonstrating that robustness can be improved without major loss in performance. These results highlight the potential for transparent and dependable AI-driven cybersecurity solutions, particularly in adversarial settings where evasion is common. By emphasizing explainability and robustness, this work contributes towards models that balance performance with trust in evolving threat landscapes.
期刊介绍:
Journal of Information Security and Applications (JISA) focuses on the original research and practice-driven applications with relevance to information security and applications. JISA provides a common linkage between a vibrant scientific and research community and industry professionals by offering a clear view on modern problems and challenges in information security, as well as identifying promising scientific and "best-practice" solutions. JISA issues offer a balance between original research work and innovative industrial approaches by internationally renowned information security experts and researchers.