{"title":"Learning an Interpretable Learning Rate Schedule via the Option Framework","authors":"Chaojing Yao","doi":"10.1109/ICTAI56018.2022.00111","DOIUrl":null,"url":null,"abstract":"Learning rate is a common and important hyperparameter in many gradient-based optimizers which are used for training machine learning models. Heuristic handcrafted learning rate schedules (LRSs) can work in many practical situations, but their design and tuning is a tedious work, and there is no guarantee that a given handcrafted LRS matches a given problem. Many works have been dedicated to automatically learning an LRS from the training dynamics of the optimization problem, but most of them share a common deficit: they borrow the algorithms designed elsewhere as methods for automatic outer-training, but those methods often lack interpretability in the context of learning an LRS. In this paper, we leverage the option framework, a generalization to the common rein-forcement learning framework, to automatically learn an LRS based on the dynamics of optimization, which takes the idea of temporal abstraction as an underlying interpretation. To meet the requirements of LLRS, the RL state is designed as consisting of the global state and the per-parameter state. We propose a policy architecture which processes these two parts according to their respective structures, and combines them to yield the input for functional computation. Experiments are carried out on classic machine learning tasks and test functions for numerical optimization to demonstrate the effectiveness and the interpretability of our method.","PeriodicalId":354314,"journal":{"name":"2022 IEEE 34th International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 34th International Conference on Tools with Artificial Intelligence (ICTAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTAI56018.2022.00111","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Learning rate is a common and important hyperparameter in many gradient-based optimizers which are used for training machine learning models. Heuristic handcrafted learning rate schedules (LRSs) can work in many practical situations, but their design and tuning is a tedious work, and there is no guarantee that a given handcrafted LRS matches a given problem. Many works have been dedicated to automatically learning an LRS from the training dynamics of the optimization problem, but most of them share a common deficit: they borrow the algorithms designed elsewhere as methods for automatic outer-training, but those methods often lack interpretability in the context of learning an LRS. In this paper, we leverage the option framework, a generalization to the common rein-forcement learning framework, to automatically learn an LRS based on the dynamics of optimization, which takes the idea of temporal abstraction as an underlying interpretation. To meet the requirements of LLRS, the RL state is designed as consisting of the global state and the per-parameter state. We propose a policy architecture which processes these two parts according to their respective structures, and combines them to yield the input for functional computation. Experiments are carried out on classic machine learning tasks and test functions for numerical optimization to demonstrate the effectiveness and the interpretability of our method.