{"title":"基于线性规划的超局部寻优超参数","authors":"Ankur Sinha , Satender Gunwal","doi":"10.1016/j.orl.2025.107287","DOIUrl":null,"url":null,"abstract":"<div><div>We introduce a linear programming-based approach for hyperparameter tuning of machine learning models. The approach finetunes continuous hyperparameters and model parameters through a linear program, enhancing model generalization in the vicinity of an initial model. The proposed method converts hyperparameter optimization into a bilevel program and identifies a descent direction to improve validation loss. The results demonstrate improvements in most cases across regression, machine learning, and deep learning tasks, with test performance enhancements ranging from 0.3% to 28.1%.</div></div>","PeriodicalId":54682,"journal":{"name":"Operations Research Letters","volume":"61 ","pages":"Article 107287"},"PeriodicalIF":0.9000,"publicationDate":"2025-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A linear programming-based hyper local search for tuning hyperparameters\",\"authors\":\"Ankur Sinha , Satender Gunwal\",\"doi\":\"10.1016/j.orl.2025.107287\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>We introduce a linear programming-based approach for hyperparameter tuning of machine learning models. The approach finetunes continuous hyperparameters and model parameters through a linear program, enhancing model generalization in the vicinity of an initial model. The proposed method converts hyperparameter optimization into a bilevel program and identifies a descent direction to improve validation loss. The results demonstrate improvements in most cases across regression, machine learning, and deep learning tasks, with test performance enhancements ranging from 0.3% to 28.1%.</div></div>\",\"PeriodicalId\":54682,\"journal\":{\"name\":\"Operations Research Letters\",\"volume\":\"61 \",\"pages\":\"Article 107287\"},\"PeriodicalIF\":0.9000,\"publicationDate\":\"2025-04-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Operations Research Letters\",\"FirstCategoryId\":\"91\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167637725000483\",\"RegionNum\":4,\"RegionCategory\":\"管理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"OPERATIONS RESEARCH & MANAGEMENT SCIENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Operations Research Letters","FirstCategoryId":"91","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167637725000483","RegionNum":4,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"OPERATIONS RESEARCH & MANAGEMENT SCIENCE","Score":null,"Total":0}
A linear programming-based hyper local search for tuning hyperparameters
We introduce a linear programming-based approach for hyperparameter tuning of machine learning models. The approach finetunes continuous hyperparameters and model parameters through a linear program, enhancing model generalization in the vicinity of an initial model. The proposed method converts hyperparameter optimization into a bilevel program and identifies a descent direction to improve validation loss. The results demonstrate improvements in most cases across regression, machine learning, and deep learning tasks, with test performance enhancements ranging from 0.3% to 28.1%.
期刊介绍:
Operations Research Letters is committed to the rapid review and fast publication of short articles on all aspects of operations research and analytics. Apart from a limitation to eight journal pages, quality, originality, relevance and clarity are the only criteria for selecting the papers to be published. ORL covers the broad field of optimization, stochastic models and game theory. Specific areas of interest include networks, routing, location, queueing, scheduling, inventory, reliability, and financial engineering. We wish to explore interfaces with other fields such as life sciences and health care, artificial intelligence and machine learning, energy distribution, and computational social sciences and humanities. Our traditional strength is in methodology, including theory, modelling, algorithms and computational studies. We also welcome novel applications and concise literature reviews.