Forensically Classifying Files Using HSOM Algorithms

G. Pierris, S. Vidalis
{"title":"Forensically Classifying Files Using HSOM Algorithms","authors":"G. Pierris, S. Vidalis","doi":"10.1109/EIDWT.2012.46","DOIUrl":null,"url":null,"abstract":"It has been accepted by Cloud Computing vendors that retrieving data from a cloud environment once they have been deleted is next to impossible. This constitutes a major hurdle for the digital forensics examiner as it greatly limits the pool of potential evidence that could be collected during an investigation. In this concept paper we will discuss a different approach to the above problem that spans across two different worlds: the world of digital forensics and the world of artificial intelligence. Block-based hash analysis works by calculating a hash value for each block of the target file that would be allocated a sector or cluster to store its data. The block hashes are then stored in a \"map\" file. The examiner then searches secondary memory areas to see if they contain blocks matching those contained in the \"map\" files. The examiner then has the ability to rebuild any file whose blocks have been located. The processes of hash-map calculation and analysis in the case of graphic images is accomplished using a single, dual-purpose EnScript in EnCase. Where a suspect file has been partially but not completely located the script will produce a PNG graphic showing exactly which blocks of the graphic have been located. This technique is extremely time and processor intensive, and does not work for unknown broken files. We hypothesize that we can use Hierarchical Self-Organizing Map algorithms in order to classify broken chains of previously unknown files, and in the future reconstruct them in order to be examined by the digital forensic examiner using the block-based hash analysis technique.","PeriodicalId":222292,"journal":{"name":"2012 Third International Conference on Emerging Intelligent Data and Web Technologies","volume":"23 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 Third International Conference on Emerging Intelligent Data and Web Technologies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/EIDWT.2012.46","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

It has been accepted by Cloud Computing vendors that retrieving data from a cloud environment once they have been deleted is next to impossible. This constitutes a major hurdle for the digital forensics examiner as it greatly limits the pool of potential evidence that could be collected during an investigation. In this concept paper we will discuss a different approach to the above problem that spans across two different worlds: the world of digital forensics and the world of artificial intelligence. Block-based hash analysis works by calculating a hash value for each block of the target file that would be allocated a sector or cluster to store its data. The block hashes are then stored in a "map" file. The examiner then searches secondary memory areas to see if they contain blocks matching those contained in the "map" files. The examiner then has the ability to rebuild any file whose blocks have been located. The processes of hash-map calculation and analysis in the case of graphic images is accomplished using a single, dual-purpose EnScript in EnCase. Where a suspect file has been partially but not completely located the script will produce a PNG graphic showing exactly which blocks of the graphic have been located. This technique is extremely time and processor intensive, and does not work for unknown broken files. We hypothesize that we can use Hierarchical Self-Organizing Map algorithms in order to classify broken chains of previously unknown files, and in the future reconstruct them in order to be examined by the digital forensic examiner using the block-based hash analysis technique.
使用HSOM算法对文件进行法医分类
云计算供应商已经接受,一旦数据被删除,从云环境中检索数据几乎是不可能的。这对数字取证官来说是一个很大的障碍,因为它极大地限制了调查过程中可以收集的潜在证据的范围。在这篇概念论文中,我们将讨论跨越两个不同世界的解决上述问题的不同方法:数字取证世界和人工智能世界。基于块的哈希分析通过计算目标文件的每个块的哈希值来工作,这些块将被分配一个扇区或集群来存储其数据。然后将块散列存储在“map”文件中。然后,考官搜索次级记忆区域,看看它们是否包含与“地图”文件中包含的块相匹配的块。然后,审查员有能力重建其块已定位的任何文件。在图形图像的情况下,哈希映射的计算和分析过程是使用EnCase中的一个单一的、双重用途的EnScript来完成的。如果可疑文件已部分定位但未完全定位,则脚本将生成PNG图形,显示已定位的图形块。这种技术非常耗时和处理器密集,并且不适用于未知的损坏文件。我们假设我们可以使用分层自组织映射算法来对先前未知文件的断裂链进行分类,并在未来重建它们,以便使用基于块的哈希分析技术进行数字法医检查。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信