基于神经网络的寻优缓存预测模型研究

Songchok Khakhaeng, C. Phongpensri
{"title":"基于神经网络的寻优缓存预测模型研究","authors":"Songchok Khakhaeng, C. Phongpensri","doi":"10.1109/KST.2016.7440514","DOIUrl":null,"url":null,"abstract":"Cache is an important component in computer architecture. It is a kind of memories that is located very close to CPU and can be accessed fast. With the current technology, the cache price is still expensive and thus the size cannot be very large. With the proper selection of the cache design, one can save the total hardware cost while certainly gaining the fast execution of programs. In this paper, we apply the data mining technique to find the proper cache model. We particularly consider to predict the cache block size. First, how to collect the address reference traces is described. Several tools are used to collect the traces. We are interested in the trace of the data mining benchmark, NUMineBench[4]. From the traces, the reference patterns are analyzed. Patterns related to block size are extracted. These features together with the trace behaviors from the simulation are used to build the prediction model, i.e, neural network. The methodology is found to be effective and can be expanded to consider other parameters such as cache capacity, associativity etc.","PeriodicalId":350687,"journal":{"name":"2016 8th International Conference on Knowledge and Smart Technology (KST)","volume":"425 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"9","resultStr":"{\"title\":\"On the finding proper cache prediction model using neural network\",\"authors\":\"Songchok Khakhaeng, C. Phongpensri\",\"doi\":\"10.1109/KST.2016.7440514\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Cache is an important component in computer architecture. It is a kind of memories that is located very close to CPU and can be accessed fast. With the current technology, the cache price is still expensive and thus the size cannot be very large. With the proper selection of the cache design, one can save the total hardware cost while certainly gaining the fast execution of programs. In this paper, we apply the data mining technique to find the proper cache model. We particularly consider to predict the cache block size. First, how to collect the address reference traces is described. Several tools are used to collect the traces. We are interested in the trace of the data mining benchmark, NUMineBench[4]. From the traces, the reference patterns are analyzed. Patterns related to block size are extracted. These features together with the trace behaviors from the simulation are used to build the prediction model, i.e, neural network. The methodology is found to be effective and can be expanded to consider other parameters such as cache capacity, associativity etc.\",\"PeriodicalId\":350687,\"journal\":{\"name\":\"2016 8th International Conference on Knowledge and Smart Technology (KST)\",\"volume\":\"425 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-02-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"9\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 8th International Conference on Knowledge and Smart Technology (KST)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/KST.2016.7440514\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 8th International Conference on Knowledge and Smart Technology (KST)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/KST.2016.7440514","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 9

摘要

高速缓存是计算机体系结构的重要组成部分。它是一种存储器,位于非常接近CPU,可以快速访问。以目前的技术,高速缓存的价格仍然昂贵,因此尺寸不能很大。通过正确选择缓存设计,可以节省总硬件成本,同时确保程序的快速执行。在本文中,我们应用数据挖掘技术来寻找合适的缓存模型。我们特别考虑预测缓存块大小。首先,描述了如何收集地址参考轨迹。有几个工具被用来收集轨迹。我们对数据挖掘基准NUMineBench[4]的跟踪很感兴趣。根据轨迹分析了参考模式。提取与块大小相关的模式。这些特征与仿真的跟踪行为一起用于构建预测模型,即神经网络。该方法被证明是有效的,并且可以扩展到考虑其他参数,如缓存容量、关联性等。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
On the finding proper cache prediction model using neural network
Cache is an important component in computer architecture. It is a kind of memories that is located very close to CPU and can be accessed fast. With the current technology, the cache price is still expensive and thus the size cannot be very large. With the proper selection of the cache design, one can save the total hardware cost while certainly gaining the fast execution of programs. In this paper, we apply the data mining technique to find the proper cache model. We particularly consider to predict the cache block size. First, how to collect the address reference traces is described. Several tools are used to collect the traces. We are interested in the trace of the data mining benchmark, NUMineBench[4]. From the traces, the reference patterns are analyzed. Patterns related to block size are extracted. These features together with the trace behaviors from the simulation are used to build the prediction model, i.e, neural network. The methodology is found to be effective and can be expanded to consider other parameters such as cache capacity, associativity etc.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信