Imrus Salehin , Md. Shamiul Islam , Pritom Saha , S.M. Noman , Azra Tuni , Md. Mehedi Hasan , Md. Abu Baten
{"title":"AutoML:利用神经架构搜索自动机器学习的系统综述","authors":"Imrus Salehin , Md. Shamiul Islam , Pritom Saha , S.M. Noman , Azra Tuni , Md. Mehedi Hasan , Md. Abu Baten","doi":"10.1016/j.jiixd.2023.10.002","DOIUrl":null,"url":null,"abstract":"<div><p>AutoML (Automated Machine Learning) is an emerging field that aims to automate the process of building machine learning models. AutoML emerged to increase productivity and efficiency by automating as much as possible the inefficient work that occurs while repeating this process whenever machine learning is applied. In particular, research has been conducted for a long time on technologies that can effectively develop high-quality models by minimizing the intervention of model developers in the process from data preprocessing to algorithm selection and tuning. In this semantic review research, we summarize the data processing requirements for AutoML approaches and provide a detailed explanation. We place greater emphasis on neural architecture search (NAS) as it currently represents a highly popular sub-topic within the field of AutoML. NAS methods use machine learning algorithms to search through a large space of possible architectures and find the one that performs best on a given task. We provide a summary of the performance achieved by representative NAS algorithms on the CIFAR-10, CIFAR-100, ImageNet and well-known benchmark datasets. Additionally, we delve into several noteworthy research directions in NAS methods including one/two-stage NAS, one-shot NAS and joint hyperparameter with architecture optimization. We discussed how the search space size and complexity in NAS can vary depending on the specific problem being addressed. To conclude, we examine several open problems (SOTA problems) within current AutoML methods that assure further investigation in future research.</p></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 1","pages":"Pages 52-81"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949715923000604/pdfft?md5=a79f7fb3cdab55edd3b7838063f99f50&pid=1-s2.0-S2949715923000604-main.pdf","citationCount":"0","resultStr":"{\"title\":\"AutoML: A systematic review on automated machine learning with neural architecture search\",\"authors\":\"Imrus Salehin , Md. Shamiul Islam , Pritom Saha , S.M. Noman , Azra Tuni , Md. Mehedi Hasan , Md. Abu Baten\",\"doi\":\"10.1016/j.jiixd.2023.10.002\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>AutoML (Automated Machine Learning) is an emerging field that aims to automate the process of building machine learning models. AutoML emerged to increase productivity and efficiency by automating as much as possible the inefficient work that occurs while repeating this process whenever machine learning is applied. In particular, research has been conducted for a long time on technologies that can effectively develop high-quality models by minimizing the intervention of model developers in the process from data preprocessing to algorithm selection and tuning. In this semantic review research, we summarize the data processing requirements for AutoML approaches and provide a detailed explanation. We place greater emphasis on neural architecture search (NAS) as it currently represents a highly popular sub-topic within the field of AutoML. NAS methods use machine learning algorithms to search through a large space of possible architectures and find the one that performs best on a given task. We provide a summary of the performance achieved by representative NAS algorithms on the CIFAR-10, CIFAR-100, ImageNet and well-known benchmark datasets. Additionally, we delve into several noteworthy research directions in NAS methods including one/two-stage NAS, one-shot NAS and joint hyperparameter with architecture optimization. We discussed how the search space size and complexity in NAS can vary depending on the specific problem being addressed. To conclude, we examine several open problems (SOTA problems) within current AutoML methods that assure further investigation in future research.</p></div>\",\"PeriodicalId\":100790,\"journal\":{\"name\":\"Journal of Information and Intelligence\",\"volume\":\"2 1\",\"pages\":\"Pages 52-81\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.sciencedirect.com/science/article/pii/S2949715923000604/pdfft?md5=a79f7fb3cdab55edd3b7838063f99f50&pid=1-s2.0-S2949715923000604-main.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Information and Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2949715923000604\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Information and Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2949715923000604","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
AutoML(自动化机器学习)是一个新兴领域,旨在实现机器学习模型构建过程的自动化。AutoML 的出现是为了尽可能自动化重复机器学习过程中出现的低效工作,从而提高生产率和效率。特别是,从数据预处理到算法选择和调整,模型开发人员在这一过程中的干预降到最低,从而有效开发出高质量模型的技术已经研究了很长时间。在这项语义回顾研究中,我们总结了 AutoML 方法的数据处理要求,并提供了详细的解释。我们更加重视神经架构搜索(NAS),因为它是目前 AutoML 领域非常热门的子课题。NAS 方法使用机器学习算法在大量可能的架构中进行搜索,找出在给定任务中表现最佳的架构。我们总结了具有代表性的 NAS 算法在 CIFAR-10、CIFAR-100、ImageNet 和知名基准数据集上取得的性能。此外,我们还深入探讨了 NAS 方法中几个值得关注的研究方向,包括单/两阶段 NAS、单次 NAS 和联合超参数与架构优化。我们讨论了 NAS 的搜索空间大小和复杂性如何因所解决的具体问题而异。最后,我们探讨了当前 AutoML 方法中的几个开放问题(SOTA 问题),这些问题值得在未来的研究中进一步探讨。
AutoML: A systematic review on automated machine learning with neural architecture search
AutoML (Automated Machine Learning) is an emerging field that aims to automate the process of building machine learning models. AutoML emerged to increase productivity and efficiency by automating as much as possible the inefficient work that occurs while repeating this process whenever machine learning is applied. In particular, research has been conducted for a long time on technologies that can effectively develop high-quality models by minimizing the intervention of model developers in the process from data preprocessing to algorithm selection and tuning. In this semantic review research, we summarize the data processing requirements for AutoML approaches and provide a detailed explanation. We place greater emphasis on neural architecture search (NAS) as it currently represents a highly popular sub-topic within the field of AutoML. NAS methods use machine learning algorithms to search through a large space of possible architectures and find the one that performs best on a given task. We provide a summary of the performance achieved by representative NAS algorithms on the CIFAR-10, CIFAR-100, ImageNet and well-known benchmark datasets. Additionally, we delve into several noteworthy research directions in NAS methods including one/two-stage NAS, one-shot NAS and joint hyperparameter with architecture optimization. We discussed how the search space size and complexity in NAS can vary depending on the specific problem being addressed. To conclude, we examine several open problems (SOTA problems) within current AutoML methods that assure further investigation in future research.