{"title":"Rethinking density ratio estimation based hyper-parameter optimization","authors":"Zi-En Fan, Feng Lian, Xin-Ran Li","doi":"10.1016/j.neunet.2024.106917","DOIUrl":null,"url":null,"abstract":"<div><div>Hyper-parameter optimization (HPO) aims to improve the performance of machine learning algorithms by identifying appropriate hyper-parameters. By converting the computation of expected improvement into density-ratio estimation problems, existing works use binary classifiers to estimate these ratio and determine the next point by maximizing the class posterior probabilities. However, these methods tend to treat different points equally and ignore some important regions, because binary classifiers are unable to capture more information about search spaces and highlight important regions. In this work, we propose a hyper-parameter optimization method by estimating ratios and selecting the next point using multi-class classifiers. First, we divide all samples into multiple classes and train multi-class classifiers. The decision boundaries of the trained classifiers allow for a finer partitioning of search spaces, offering richer insights into the distribution of hyper-parameters within search spaces. We then define an acquisition function as a weighted sum of multi-class classifiers’ outputs, with these weights determined by samples in each class. By assigning different weights to each class posterior probability in our acquisition function, points within search spaces are no longer treated equally. Experimental results on three representative tasks demonstrate that our method achieves a significant improvement in immediate regrets and convergence speed.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"182 ","pages":"Article 106917"},"PeriodicalIF":6.0000,"publicationDate":"2024-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0893608024008463","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Hyper-parameter optimization (HPO) aims to improve the performance of machine learning algorithms by identifying appropriate hyper-parameters. By converting the computation of expected improvement into density-ratio estimation problems, existing works use binary classifiers to estimate these ratio and determine the next point by maximizing the class posterior probabilities. However, these methods tend to treat different points equally and ignore some important regions, because binary classifiers are unable to capture more information about search spaces and highlight important regions. In this work, we propose a hyper-parameter optimization method by estimating ratios and selecting the next point using multi-class classifiers. First, we divide all samples into multiple classes and train multi-class classifiers. The decision boundaries of the trained classifiers allow for a finer partitioning of search spaces, offering richer insights into the distribution of hyper-parameters within search spaces. We then define an acquisition function as a weighted sum of multi-class classifiers’ outputs, with these weights determined by samples in each class. By assigning different weights to each class posterior probability in our acquisition function, points within search spaces are no longer treated equally. Experimental results on three representative tasks demonstrate that our method achieves a significant improvement in immediate regrets and convergence speed.
期刊介绍:
Neural Networks is a platform that aims to foster an international community of scholars and practitioners interested in neural networks, deep learning, and other approaches to artificial intelligence and machine learning. Our journal invites submissions covering various aspects of neural networks research, from computational neuroscience and cognitive modeling to mathematical analyses and engineering applications. By providing a forum for interdisciplinary discussions between biology and technology, we aim to encourage the development of biologically-inspired artificial intelligence.