Sparse DARTS with Various Recovery Algorithms

Yanqing Hu, Qing Ye, Huan-Shuan Fu, Jiancheng Lv
{"title":"Sparse DARTS with Various Recovery Algorithms","authors":"Yanqing Hu, Qing Ye, Huan-Shuan Fu, Jiancheng Lv","doi":"10.1145/3532213.3532225","DOIUrl":null,"url":null,"abstract":"Designing an efficient neural architecture search method is an open and challenging problem over the last few years. A typical and well-performed strategy is gradient-based methods (i.e., Differentiable Architecture Search (DARTS)), which mainly searches the target sparse child graph from a trainable dense super graph. However, during the searching phrase, training the dense super graph usually requires excessively computational resources. Besides, the training based on a dense graph is excessively inefficient, and the memory consumption is prohibitively high. To alleviate this shortcoming, recently Iterative Shrinkage Thresholding Algorithm (ISTA), a sparse coding recovery algorithm, has been applied to DARTS, which directly optimizes the compressed representation of the super graph, and saves the memory and time consumption. Indeed, there are several such kinds of sparse coding recovery algorithms, and ISTA is not the best one in terms of recovery efficiency and effectiveness. To investigate the impact of different sparse coding recovery algorithm on performance in DARTS and provide some insights. Firstly, we design several sparse DARTS based on different sparse coding recovery algorithms (i.e., LISTA, CoD, and Lars). Then a series of controlled experiments on selected algorithms are conducted. The accuracy, search time and other indicators of the model are collected and compared. Sufficient theoretical analysis and experimental exploration reveal that the different compression algorithms show different characteristics on the sparse DARTS. Specifically, Lars-NAS tends to choose the operation with fewer parameters, while Cod-NAS is the simplest of the four recovery algorithms, and its consuming time is very short, but the CoD-NAS model is unstable. Particularly, LISTA-NAS achieves the accurate results with stable recovery time. Thus, it can be seen that all compression algorithms are available to utilized according to different environments and requirements.","PeriodicalId":333199,"journal":{"name":"Proceedings of the 8th International Conference on Computing and Artificial Intelligence","volume":"67 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 8th International Conference on Computing and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3532213.3532225","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Designing an efficient neural architecture search method is an open and challenging problem over the last few years. A typical and well-performed strategy is gradient-based methods (i.e., Differentiable Architecture Search (DARTS)), which mainly searches the target sparse child graph from a trainable dense super graph. However, during the searching phrase, training the dense super graph usually requires excessively computational resources. Besides, the training based on a dense graph is excessively inefficient, and the memory consumption is prohibitively high. To alleviate this shortcoming, recently Iterative Shrinkage Thresholding Algorithm (ISTA), a sparse coding recovery algorithm, has been applied to DARTS, which directly optimizes the compressed representation of the super graph, and saves the memory and time consumption. Indeed, there are several such kinds of sparse coding recovery algorithms, and ISTA is not the best one in terms of recovery efficiency and effectiveness. To investigate the impact of different sparse coding recovery algorithm on performance in DARTS and provide some insights. Firstly, we design several sparse DARTS based on different sparse coding recovery algorithms (i.e., LISTA, CoD, and Lars). Then a series of controlled experiments on selected algorithms are conducted. The accuracy, search time and other indicators of the model are collected and compared. Sufficient theoretical analysis and experimental exploration reveal that the different compression algorithms show different characteristics on the sparse DARTS. Specifically, Lars-NAS tends to choose the operation with fewer parameters, while Cod-NAS is the simplest of the four recovery algorithms, and its consuming time is very short, but the CoD-NAS model is unstable. Particularly, LISTA-NAS achieves the accurate results with stable recovery time. Thus, it can be seen that all compression algorithms are available to utilized according to different environments and requirements.
各种恢复算法的稀疏dart
设计一种高效的神经结构搜索方法是近年来一个开放性和挑战性的问题。一种典型且性能良好的策略是基于梯度的方法(即可微分架构搜索(DARTS)),它主要从可训练的密集超图中搜索目标稀疏子图。然而,在搜索阶段,训练密集超图通常需要耗费过多的计算资源。此外,基于密集图的训练效率非常低,并且内存消耗过高。为了克服这一缺点,近年来将稀疏编码恢复算法迭代收缩阈值算法(Iterative Shrinkage threshold Algorithm, ISTA)应用于dart,该算法直接优化了超级图的压缩表示,节省了内存和时间消耗。事实上,这样的稀疏编码恢复算法有很多种,从恢复效率和效果来看,ISTA并不是最好的一种。研究不同稀疏编码恢复算法对dart中性能的影响,并提供一些见解。首先,我们基于不同的稀疏编码恢复算法(LISTA、CoD和Lars)设计了几种稀疏dart。然后对选定的算法进行了一系列的对照实验。收集并比较了模型的准确率、搜索时间等指标。充分的理论分析和实验探索表明,不同的压缩算法在稀疏型dart上表现出不同的特点。具体而言,Lars-NAS倾向于选择参数较少的操作,而Cod-NAS是四种恢复算法中最简单的,其消耗时间非常短,但Cod-NAS模型不稳定。其中,LISTA-NAS的检测结果准确,恢复时间稳定。由此可见,所有的压缩算法都是可以根据不同的环境和需求来使用的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信