DP-SWAP:基于动态规划的快速交换策略

IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS
Weiduo Chen , Xiaoshe Dong , Qiang Wang
{"title":"DP-SWAP:基于动态规划的快速交换策略","authors":"Weiduo Chen ,&nbsp;Xiaoshe Dong ,&nbsp;Qiang Wang","doi":"10.1016/j.future.2025.108071","DOIUrl":null,"url":null,"abstract":"<div><div>Neural Architecture Search (NAS) has emerged as an effective approach for automating neural network design. However, NAS imposes significant GPU memory pressure due to the need to evaluate numerous candidate models during training. While tensor swapping helps reduce memory usage, existing tensor selection methods rely on extensive iterative searches, which require repeatedly traversing model computation graphs to evaluate the impact of swapping schemes–leading to high time complexity and poor scalability in dynamic NAS scenarios.</div><div>To address this issue, we propose DP-SWAP, a novel tensor swapping strategy based on dynamic programming. By leveraging the optimal substructure property of the tensor selection problem, DP-SWAP computes effective swapping schemes with only <span><math><mrow><mi>O</mi><mo>(</mo><mi>n</mi><mo>)</mo></mrow></math></span> time complexity, allows for fast and adaptive decision-making during NAS model exploration.</div><div>Experimental results show that DP-SWAP achieves training performance comparable to state-of-the-art heuristic methods, while reducing swapping decision time by over 3 orders of magnitude, thus effectively alleviating GPU memory bottlenecks in NAS.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"175 ","pages":"Article 108071"},"PeriodicalIF":6.2000,"publicationDate":"2025-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"DP-SWAP: Fast Swapping Strategy Based on Dynamic Programming\",\"authors\":\"Weiduo Chen ,&nbsp;Xiaoshe Dong ,&nbsp;Qiang Wang\",\"doi\":\"10.1016/j.future.2025.108071\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Neural Architecture Search (NAS) has emerged as an effective approach for automating neural network design. However, NAS imposes significant GPU memory pressure due to the need to evaluate numerous candidate models during training. While tensor swapping helps reduce memory usage, existing tensor selection methods rely on extensive iterative searches, which require repeatedly traversing model computation graphs to evaluate the impact of swapping schemes–leading to high time complexity and poor scalability in dynamic NAS scenarios.</div><div>To address this issue, we propose DP-SWAP, a novel tensor swapping strategy based on dynamic programming. By leveraging the optimal substructure property of the tensor selection problem, DP-SWAP computes effective swapping schemes with only <span><math><mrow><mi>O</mi><mo>(</mo><mi>n</mi><mo>)</mo></mrow></math></span> time complexity, allows for fast and adaptive decision-making during NAS model exploration.</div><div>Experimental results show that DP-SWAP achieves training performance comparable to state-of-the-art heuristic methods, while reducing swapping decision time by over 3 orders of magnitude, thus effectively alleviating GPU memory bottlenecks in NAS.</div></div>\",\"PeriodicalId\":55132,\"journal\":{\"name\":\"Future Generation Computer Systems-The International Journal of Escience\",\"volume\":\"175 \",\"pages\":\"Article 108071\"},\"PeriodicalIF\":6.2000,\"publicationDate\":\"2025-08-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Future Generation Computer Systems-The International Journal of Escience\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167739X25003656\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, THEORY & METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X25003656","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0

摘要

神经结构搜索(NAS)作为神经网络自动化设计的一种有效方法而出现。然而,由于需要在训练期间评估众多候选模型,NAS施加了显著的GPU内存压力。虽然张量交换有助于减少内存使用,但现有的张量选择方法依赖于大量的迭代搜索,这需要反复遍历模型计算图来评估交换方案的影响,从而导致动态NAS场景中的高时间复杂度和差可扩展性。为了解决这个问题,我们提出了一种新的基于动态规划的张量交换策略DP-SWAP。通过利用张量选择问题的最优子结构特性,DP-SWAP计算有效的交换方案,只有O(n)时间复杂度,允许在NAS模型探索过程中快速和自适应决策。实验结果表明,DP-SWAP的训练性能与最先进的启发式方法相当,同时将交换决策时间减少了3个数量级以上,从而有效缓解了NAS中GPU内存瓶颈。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
DP-SWAP: Fast Swapping Strategy Based on Dynamic Programming
Neural Architecture Search (NAS) has emerged as an effective approach for automating neural network design. However, NAS imposes significant GPU memory pressure due to the need to evaluate numerous candidate models during training. While tensor swapping helps reduce memory usage, existing tensor selection methods rely on extensive iterative searches, which require repeatedly traversing model computation graphs to evaluate the impact of swapping schemes–leading to high time complexity and poor scalability in dynamic NAS scenarios.
To address this issue, we propose DP-SWAP, a novel tensor swapping strategy based on dynamic programming. By leveraging the optimal substructure property of the tensor selection problem, DP-SWAP computes effective swapping schemes with only O(n) time complexity, allows for fast and adaptive decision-making during NAS model exploration.
Experimental results show that DP-SWAP achieves training performance comparable to state-of-the-art heuristic methods, while reducing swapping decision time by over 3 orders of magnitude, thus effectively alleviating GPU memory bottlenecks in NAS.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
19.90
自引率
2.70%
发文量
376
审稿时长
10.6 months
期刊介绍: Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications. Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration. Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信