Andy Brown, Aaron Tuor, Brian Hutchinson, Nicole Nichols
{"title":"可解释系统日志异常检测的递归神经网络注意机制","authors":"Andy Brown, Aaron Tuor, Brian Hutchinson, Nicole Nichols","doi":"10.1145/3217871.3217872","DOIUrl":null,"url":null,"abstract":"Deep learning has recently demonstrated state-of-the art performance on key tasks related to the maintenance of computer systems, such as intrusion detection, denial of service attack detection, hardware and software system failures, and malware detection. In these contexts, model interpretability is vital for administrator and analyst to trust and act on the automated analysis of machine learning models. Deep learning methods have been criticized as black box oracles which allow limited insight into decision factors. In this work we seek to bridge the gap between the impressive performance of deep learning models and the need for interpretable model introspection. To this end we present recurrent neural network (RNN) language models augmented with attention for anomaly detection in system logs. Our methods are generally applicable to any computer system and logging source. By incorporating attention variants into our RNN language models we create opportunities for model introspection and analysis without sacrificing state-of-the art performance. We demonstrate model performance and illustrate model interpretability on an intrusion detection task using the Los Alamos National Laboratory (LANL) cyber security dataset, reporting upward of 0.99 area under the receiver operator characteristic curve despite being trained only on a single day's worth of data.","PeriodicalId":174025,"journal":{"name":"Proceedings of the First Workshop on Machine Learning for Computing Systems","volume":"26 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"151","resultStr":"{\"title\":\"Recurrent Neural Network Attention Mechanisms for Interpretable System Log Anomaly Detection\",\"authors\":\"Andy Brown, Aaron Tuor, Brian Hutchinson, Nicole Nichols\",\"doi\":\"10.1145/3217871.3217872\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Deep learning has recently demonstrated state-of-the art performance on key tasks related to the maintenance of computer systems, such as intrusion detection, denial of service attack detection, hardware and software system failures, and malware detection. In these contexts, model interpretability is vital for administrator and analyst to trust and act on the automated analysis of machine learning models. Deep learning methods have been criticized as black box oracles which allow limited insight into decision factors. In this work we seek to bridge the gap between the impressive performance of deep learning models and the need for interpretable model introspection. To this end we present recurrent neural network (RNN) language models augmented with attention for anomaly detection in system logs. Our methods are generally applicable to any computer system and logging source. By incorporating attention variants into our RNN language models we create opportunities for model introspection and analysis without sacrificing state-of-the art performance. We demonstrate model performance and illustrate model interpretability on an intrusion detection task using the Los Alamos National Laboratory (LANL) cyber security dataset, reporting upward of 0.99 area under the receiver operator characteristic curve despite being trained only on a single day's worth of data.\",\"PeriodicalId\":174025,\"journal\":{\"name\":\"Proceedings of the First Workshop on Machine Learning for Computing Systems\",\"volume\":\"26 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-03-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"151\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the First Workshop on Machine Learning for Computing Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3217871.3217872\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the First Workshop on Machine Learning for Computing Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3217871.3217872","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 151
摘要
深度学习最近在与计算机系统维护相关的关键任务上展示了最先进的性能,例如入侵检测、拒绝服务攻击检测、硬件和软件系统故障以及恶意软件检测。在这些环境中,模型的可解释性对于管理员和分析人员信任并对机器学习模型的自动分析采取行动至关重要。深度学习方法被批评为黑盒预言机,对决策因素的洞察力有限。在这项工作中,我们试图弥合深度学习模型令人印象深刻的性能与对可解释模型内省的需求之间的差距。为此,我们提出了递归神经网络(RNN)语言模型,增强了对系统日志异常检测的关注。我们的方法一般适用于任何计算机系统和日志源。通过将注意力变量合并到我们的RNN语言模型中,我们为模型自省和分析创造了机会,而不会牺牲最先进的性能。我们使用洛斯阿拉莫斯国家实验室(Los Alamos National Laboratory, LANL)网络安全数据集在入侵检测任务上演示了模型性能并说明了模型的可解释性,尽管仅在一天的数据上进行了训练,但在接收器操作员特征曲线下报告了0.99以上的区域。
Recurrent Neural Network Attention Mechanisms for Interpretable System Log Anomaly Detection
Deep learning has recently demonstrated state-of-the art performance on key tasks related to the maintenance of computer systems, such as intrusion detection, denial of service attack detection, hardware and software system failures, and malware detection. In these contexts, model interpretability is vital for administrator and analyst to trust and act on the automated analysis of machine learning models. Deep learning methods have been criticized as black box oracles which allow limited insight into decision factors. In this work we seek to bridge the gap between the impressive performance of deep learning models and the need for interpretable model introspection. To this end we present recurrent neural network (RNN) language models augmented with attention for anomaly detection in system logs. Our methods are generally applicable to any computer system and logging source. By incorporating attention variants into our RNN language models we create opportunities for model introspection and analysis without sacrificing state-of-the art performance. We demonstrate model performance and illustrate model interpretability on an intrusion detection task using the Los Alamos National Laboratory (LANL) cyber security dataset, reporting upward of 0.99 area under the receiver operator characteristic curve despite being trained only on a single day's worth of data.