Malware Detection for Forensic Memory Using Deep Recurrent Neural Networks

Ioannis Karamitsos, Aishwarya Afzulpurkar, T. Trafalis
{"title":"Malware Detection for Forensic Memory Using Deep Recurrent Neural Networks","authors":"Ioannis Karamitsos, Aishwarya Afzulpurkar, T. Trafalis","doi":"10.4236/jis.2020.112007","DOIUrl":null,"url":null,"abstract":"Memory forensics is a young but fast-growing area of research and a promising one for the field of computer forensics. The learned model is proposed to reside in an isolated core with strict communication restrictions to achieve incorruptibility as well as efficiency, therefore providing a probabilistic memory-level view of the system that is consistent with the user-level view. The lower level memory blocks are constructed using primary block sequences of varying sizes that are fed as input into Long-Short Term Memory (LSTM) models. Four configurations of the LSTM model are explored by adding bi- directionality as well as attention. Assembly level data from 50 Windows portable executable (PE) files are extracted, and basic blocks are constructed using the IDA Disassembler toolkit. The results show that longer primary block sequences result in richer LSTM hidden layer representations. The hidden states are fed as features into Max pooling layers or Attention layers, depending on the configuration being tested, and the final classification is performed using Logistic Regression with a single hidden layer. The bidirectional LSTM with Attention proved to be the best model, used on basic block sequences of size 29. The differences between the model’s ROC curves indicate a strong reliance on the lower level, instructional features, as opposed to metadata or string features.","PeriodicalId":57259,"journal":{"name":"信息安全(英文)","volume":"11 1","pages":"103-120"},"PeriodicalIF":0.0000,"publicationDate":"2020-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"信息安全(英文)","FirstCategoryId":"1093","ListUrlMain":"https://doi.org/10.4236/jis.2020.112007","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

Memory forensics is a young but fast-growing area of research and a promising one for the field of computer forensics. The learned model is proposed to reside in an isolated core with strict communication restrictions to achieve incorruptibility as well as efficiency, therefore providing a probabilistic memory-level view of the system that is consistent with the user-level view. The lower level memory blocks are constructed using primary block sequences of varying sizes that are fed as input into Long-Short Term Memory (LSTM) models. Four configurations of the LSTM model are explored by adding bi- directionality as well as attention. Assembly level data from 50 Windows portable executable (PE) files are extracted, and basic blocks are constructed using the IDA Disassembler toolkit. The results show that longer primary block sequences result in richer LSTM hidden layer representations. The hidden states are fed as features into Max pooling layers or Attention layers, depending on the configuration being tested, and the final classification is performed using Logistic Regression with a single hidden layer. The bidirectional LSTM with Attention proved to be the best model, used on basic block sequences of size 29. The differences between the model’s ROC curves indicate a strong reliance on the lower level, instructional features, as opposed to metadata or string features.
基于深度递归神经网络的取证记忆恶意软件检测
内存取证是一个年轻但快速发展的研究领域,也是计算机取证领域的一个有前途的领域。所学习的模型被提议驻留在具有严格通信限制的隔离核心中,以实现不可破坏性和效率,从而提供与用户级别视图一致的系统的概率内存级别视图。使用不同大小的主块序列来构建较低级别的存储器块,所述主块序列作为输入被馈送到长短期存储器(LSTM)模型中。通过添加双向性和注意力来探索LSTM模型的四种配置。从50个Windows可移植可执行文件(PE)中提取汇编级数据,并使用IDA Disassembler工具包构建基本块。结果表明,主块序列越长,LSTM隐层表示越丰富。根据测试的配置,将隐藏状态作为特征输入到最大池化层或注意力层,并使用具有单个隐藏层的逻辑回归进行最终分类。具有注意力的双向LSTM被证明是最好的模型,用于大小为29的基本块序列。模型ROC曲线之间的差异表明,与元数据或字符串特征相比,它强烈依赖于较低级别的教学特征。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
211
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信