{"title":"An effective surrogate-assisted rank method for evolutionary neural architecture search","authors":"Yu Xue, Anjing Zhu","doi":"10.1016/j.asoc.2024.112392","DOIUrl":null,"url":null,"abstract":"<div><div>Evolutionary neural architecture search (ENAS) is able to automatically design high-performed architectures of deep neural networks (DNNs) for specific tasks. In recent years, surrogate models have gained significant traction because they can estimate the performance of neural architectures, avoiding excessive computational costs for training. However, most existing surrogate models primarily predict the performance of architectures directly or predict pairwise comparison relationships, which makes it challenging to obtain the rank of a group of architectures when training samples are limited. To address this problem, we propose TCMR-ENAS, an effective triple-competition model-assisted rank method for ENAS. TCMR-ENAS employs a novel triple-competition surrogate model combined with a score-based fitness evaluation method to predict group performance rank. Moreover, a progressive online learning method is proposed to enhance the predictive performance of the triple-competition surrogate model in the framework of modified genetic search. To validate the effectiveness of TCMR-ENAS, we conducted a series of experiments on NAS-Bench-101, NAS-Bench-201, NATS-Bench and NAS-Bench-301, respectively. Experimental results show that TCMR-ENAS can achieve better performance with lower computational resources. The accuracies of searched architectures achieve the best results compared with those of the state-of-the-art methods with limited training samples. In addition, the factors that may influence the effectiveness of TCMR-ENAS are explored in the ablation studies.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":"167 ","pages":"Article 112392"},"PeriodicalIF":7.2000,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Soft Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1568494624011669","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Evolutionary neural architecture search (ENAS) is able to automatically design high-performed architectures of deep neural networks (DNNs) for specific tasks. In recent years, surrogate models have gained significant traction because they can estimate the performance of neural architectures, avoiding excessive computational costs for training. However, most existing surrogate models primarily predict the performance of architectures directly or predict pairwise comparison relationships, which makes it challenging to obtain the rank of a group of architectures when training samples are limited. To address this problem, we propose TCMR-ENAS, an effective triple-competition model-assisted rank method for ENAS. TCMR-ENAS employs a novel triple-competition surrogate model combined with a score-based fitness evaluation method to predict group performance rank. Moreover, a progressive online learning method is proposed to enhance the predictive performance of the triple-competition surrogate model in the framework of modified genetic search. To validate the effectiveness of TCMR-ENAS, we conducted a series of experiments on NAS-Bench-101, NAS-Bench-201, NATS-Bench and NAS-Bench-301, respectively. Experimental results show that TCMR-ENAS can achieve better performance with lower computational resources. The accuracies of searched architectures achieve the best results compared with those of the state-of-the-art methods with limited training samples. In addition, the factors that may influence the effectiveness of TCMR-ENAS are explored in the ablation studies.
期刊介绍:
Applied Soft Computing is an international journal promoting an integrated view of soft computing to solve real life problems.The focus is to publish the highest quality research in application and convergence of the areas of Fuzzy Logic, Neural Networks, Evolutionary Computing, Rough Sets and other similar techniques to address real world complexities.
Applied Soft Computing is a rolling publication: articles are published as soon as the editor-in-chief has accepted them. Therefore, the web site will continuously be updated with new articles and the publication time will be short.