Is one hyperparameter optimizer enough?

Huy Tu, V. Nair
{"title":"Is one hyperparameter optimizer enough?","authors":"Huy Tu, V. Nair","doi":"10.1145/3278142.3278145","DOIUrl":null,"url":null,"abstract":"Hyperparameter tuning is the black art of automatically finding a good combination of control parameters for a data miner. While widely applied in empirical Software Engineering, there has not been much discussion on which hyperparameter tuner is best for software analytics.To address this gap in the literature, this paper applied a range of hyperparameter optimizers (grid search, random search, differential evolution, and Bayesian optimization) to a defect prediction problem. Surprisingly, no hyperparameter optimizer was observed to be “best” and, for one of the two evaluation measures studied here (F-measure), hyperparameter optimization, in 50% of cases, was no better than using default configurations. We conclude that hyperparameter optimization is more nuanced than previously believed. While such optimization can certainly lead to large improvements in the performance of classifiers used in software analytics, it remains to be seen which specific optimizers should be applied to a new dataset.","PeriodicalId":108238,"journal":{"name":"Proceedings of the 4th ACM SIGSOFT International Workshop on Software Analytics","volume":"3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"15","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 4th ACM SIGSOFT International Workshop on Software Analytics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3278142.3278145","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 15

Abstract

Hyperparameter tuning is the black art of automatically finding a good combination of control parameters for a data miner. While widely applied in empirical Software Engineering, there has not been much discussion on which hyperparameter tuner is best for software analytics.To address this gap in the literature, this paper applied a range of hyperparameter optimizers (grid search, random search, differential evolution, and Bayesian optimization) to a defect prediction problem. Surprisingly, no hyperparameter optimizer was observed to be “best” and, for one of the two evaluation measures studied here (F-measure), hyperparameter optimization, in 50% of cases, was no better than using default configurations. We conclude that hyperparameter optimization is more nuanced than previously believed. While such optimization can certainly lead to large improvements in the performance of classifiers used in software analytics, it remains to be seen which specific optimizers should be applied to a new dataset.
一个超参数优化器就足够了吗?
超参数调优是自动为数据挖掘器找到控制参数的良好组合的黑色艺术。尽管超参数调谐器在经验软件工程中得到了广泛的应用,但对于哪种超参数调谐器最适合软件分析,并没有太多的讨论。为了解决文献中的这一空白,本文应用了一系列超参数优化器(网格搜索、随机搜索、差分进化和贝叶斯优化)来解决缺陷预测问题。令人惊讶的是,没有超参数优化器被观察到是“最佳的”,并且,对于这里研究的两个评估度量之一(f度量),在50%的情况下,超参数优化并不比使用默认配置更好。我们得出结论,超参数优化比以前认为的更微妙。虽然这样的优化肯定会大大提高软件分析中使用的分类器的性能,但对于新数据集应该应用哪些特定的优化器还有待观察。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信