网络应用程序中的直接内存访问使用优化,以减少内存延迟和能耗

A. Bartzas, Miguel Peón Quirós, S. Mamagkakis, F. Catthoor, D. Soudris, J. Mendias
{"title":"网络应用程序中的直接内存访问使用优化,以减少内存延迟和能耗","authors":"A. Bartzas, Miguel Peón Quirós, S. Mamagkakis, F. Catthoor, D. Soudris, J. Mendias","doi":"10.3233/JEC-2009-0096","DOIUrl":null,"url":null,"abstract":"Today, wireless networks are becoming increasingly ubiquitous. Usually several complex multi-threaded applications are mapped on a single embedded system and each of them is triggered by a different input stream (in accordance with the run-time behaviours of the user and the environment). This dynamicity renders the task of fully analyzing at design-time these systems very complex, if not impossible. Therefore, run-time information has to be used in order to produce an efficient design. This introduces new challenges, especially for embedded system designers using a Direct Memory Access (DMA) module, who have to know in advance the memory transfer behaviour of the whole system, in order to design and program their DMA efficiently. This is especially important in embedded systems with DRAM memories as the concurrent accesses from different processing elements can adversely affect the page-based architecture of these memory elements. Even more, the increasingly common usage of dynamic data types further complicates the problem because the exact location of data instances in the memory is unknown at design-time. In this paper we propose a system-level optimization methodology to adapt the DMA usage parameters automatically at run-time, according to online information. With our proposed optimization approach we manage to reduce the mean latency of the memory transfers by more than 18%, thus reducing the average number of cycles that processing elements or DMAs have to waste waiting for data from the main memory, while optimizing energy consumption and system responsiveness. We evaluate our approach using a set of real-life applications and real wireless dynamic streams.","PeriodicalId":422048,"journal":{"name":"J. Embed. Comput.","volume":"79 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2009-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Direct memory access usage optimization in network applications for reduced memory latency and energy consumption\",\"authors\":\"A. Bartzas, Miguel Peón Quirós, S. Mamagkakis, F. Catthoor, D. Soudris, J. Mendias\",\"doi\":\"10.3233/JEC-2009-0096\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Today, wireless networks are becoming increasingly ubiquitous. Usually several complex multi-threaded applications are mapped on a single embedded system and each of them is triggered by a different input stream (in accordance with the run-time behaviours of the user and the environment). This dynamicity renders the task of fully analyzing at design-time these systems very complex, if not impossible. Therefore, run-time information has to be used in order to produce an efficient design. This introduces new challenges, especially for embedded system designers using a Direct Memory Access (DMA) module, who have to know in advance the memory transfer behaviour of the whole system, in order to design and program their DMA efficiently. This is especially important in embedded systems with DRAM memories as the concurrent accesses from different processing elements can adversely affect the page-based architecture of these memory elements. Even more, the increasingly common usage of dynamic data types further complicates the problem because the exact location of data instances in the memory is unknown at design-time. In this paper we propose a system-level optimization methodology to adapt the DMA usage parameters automatically at run-time, according to online information. With our proposed optimization approach we manage to reduce the mean latency of the memory transfers by more than 18%, thus reducing the average number of cycles that processing elements or DMAs have to waste waiting for data from the main memory, while optimizing energy consumption and system responsiveness. We evaluate our approach using a set of real-life applications and real wireless dynamic streams.\",\"PeriodicalId\":422048,\"journal\":{\"name\":\"J. Embed. Comput.\",\"volume\":\"79 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2009-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"J. Embed. Comput.\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3233/JEC-2009-0096\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"J. Embed. Comput.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3233/JEC-2009-0096","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

今天,无线网络正变得越来越普遍。通常,几个复杂的多线程应用程序被映射到一个嵌入式系统上,每个应用程序都由不同的输入流触发(根据用户和环境的运行时行为)。这种动态性使得在设计时对这些系统进行全面分析的任务变得非常复杂,如果不是不可能的话。因此,为了产生有效的设计,必须使用运行时信息。这带来了新的挑战,特别是对于使用直接内存访问(DMA)模块的嵌入式系统设计者来说,他们必须提前知道整个系统的内存传输行为,以便有效地设计和编程他们的DMA。这在具有DRAM内存的嵌入式系统中尤其重要,因为来自不同处理元素的并发访问可能对这些内存元素的基于页面的体系结构产生不利影响。而且,动态数据类型越来越普遍的使用使问题进一步复杂化,因为数据实例在内存中的确切位置在设计时是未知的。在本文中,我们提出了一种系统级优化方法,在运行时根据在线信息自动调整DMA使用参数。通过我们提出的优化方法,我们设法将内存传输的平均延迟减少了18%以上,从而减少了处理元素或dma必须浪费的平均周期数,等待主存储器的数据,同时优化了能耗和系统响应能力。我们使用一组真实的应用程序和真实的无线动态流来评估我们的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Direct memory access usage optimization in network applications for reduced memory latency and energy consumption
Today, wireless networks are becoming increasingly ubiquitous. Usually several complex multi-threaded applications are mapped on a single embedded system and each of them is triggered by a different input stream (in accordance with the run-time behaviours of the user and the environment). This dynamicity renders the task of fully analyzing at design-time these systems very complex, if not impossible. Therefore, run-time information has to be used in order to produce an efficient design. This introduces new challenges, especially for embedded system designers using a Direct Memory Access (DMA) module, who have to know in advance the memory transfer behaviour of the whole system, in order to design and program their DMA efficiently. This is especially important in embedded systems with DRAM memories as the concurrent accesses from different processing elements can adversely affect the page-based architecture of these memory elements. Even more, the increasingly common usage of dynamic data types further complicates the problem because the exact location of data instances in the memory is unknown at design-time. In this paper we propose a system-level optimization methodology to adapt the DMA usage parameters automatically at run-time, according to online information. With our proposed optimization approach we manage to reduce the mean latency of the memory transfers by more than 18%, thus reducing the average number of cycles that processing elements or DMAs have to waste waiting for data from the main memory, while optimizing energy consumption and system responsiveness. We evaluate our approach using a set of real-life applications and real wireless dynamic streams.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信