Autotune: A Derivative-free Optimization Framework for Hyperparameter Tuning

被引:60
|
作者
Koch, Patrick [1 ]
Golovidov, Oleg [1 ]
Gardner, Steven [1 ]
Wujek, Brett [1 ]
Griffin, Joshua [1 ]
Xu, Yan [1 ]
机构
[1] SAS Inst Inc, Cary, NC 27513 USA
关键词
Derivative-free Optimization; Stochastic Optimization; Bayesian Optimization; Hyperparameters; Distributed Computing System; SEARCH;
D O I
10.1145/3219819.3219837
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Machine learning applications often require hyperparameter tuning. The hyperparameters usually drive both the efficiency of the model training process and the resulting model quality. For hyperparameter tuning, machine learning algorithms are complex black-boxes. This creates a class of challenging optimization problems, whose objective functions tend to be nonsmooth, discontinuous, unpredictably varying in computational expense, and include continuous, categorical, and/or integer variables. Further, function evaluations can fail for a variety of reasons including numerical difficulties or hardware failures. Additionally, not all hyperparameter value combinations are compatible, which creates so called hidden constraints. Robust and efficient optimization algorithms are needed for hyper-parameter tuning. In this paper we present an automated parallel derivative-free optimization framework called Autotune, which combines a number of specialized sampling and search methods that are very effective in tuning machine learning models despite these challenges. Autotune provides significantly improved models over using default hyperparameter settings with minimal user interaction on real-world applications. Given the inherent expense of training numerous candidate models, we demonstrate the effectiveness of Autotune's search methods and the efficient distributed and parallel paradigms for training and tuning models, and also discuss the resource trade-offs associated with the ability to both distribute the training process and parallelize the tuning process.
引用
收藏
页码:443 / 452
页数:10
相关论文
共 50 条
  • [11] Derivative-free optimization methods
    Larson, Jeffrey
    Menickelly, Matt
    Wild, Stefan M.
    ACTA NUMERICA, 2019, 28 : 287 - 404
  • [12] Derivative-Free and Blackbox Optimization
    Huyer, W.
    MONATSHEFTE FUR MATHEMATIK, 2020, 192 (02): : 480 - 480
  • [13] Derivative-Free Optimization with Adaptive Experience for Efficient Hyper-Parameter Tuning
    Hu, Yi-Qi
    Liu, Zelin
    Yang, Hua
    Yu, Yang
    Liu, Yunfeng
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1207 - 1214
  • [14] Towards a Framework for Nonlinear Predictive Control using Derivative-Free Optimization
    McInerney, Ian
    Nita, Lucian
    Nie, Yuanbo
    Oliveri, Alberto
    Kerrigan, Eric C.
    IFAC PAPERSONLINE, 2021, 54 (06): : 284 - 289
  • [15] An Online, Derivative-Free Optimization Approach to Auto-tuning of Computing Systems
    Poojary, Sudheer
    Raghavendra, Ramya
    Manjunath, D.
    DISTRIBUTED COMPUTING AND NETWORKING, PROCEEDINGS, 2010, 5935 : 434 - 445
  • [16] ZOOpt: a toolbox for derivative-free optimization
    Liu, Yu-Ren
    Hu, Yi-Qi
    Qian, Hong
    Qian, Chao
    Yu, Yang
    SCIENCE CHINA-INFORMATION SCIENCES, 2022, 65 (10)
  • [17] ZOOpt: a toolbox for derivative-free optimization
    Yu-Ren LIU
    Yi-Qi HU
    Hong QIAN
    Chao QIAN
    Yang YU
    ScienceChina(InformationSciences), 2022, 65 (10) : 293 - 294
  • [18] ZOOpt: a toolbox for derivative-free optimization
    Yu-Ren Liu
    Yi-Qi Hu
    Hong Qian
    Chao Qian
    Yang Yu
    Science China Information Sciences, 2022, 65
  • [19] A derivative-free algorithm for unconstrained optimization
    Peng Y.
    Liu Z.
    Applied Mathematics-A Journal of Chinese Universities, 2005, 20 (4) : 491 - 498
  • [20] Derivative-Free Optimization via Classification
    Yu, Yang
    Qian, Hong
    Hu, Yi-Qi
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2286 - 2292