On `p-hyperparameter Learning via Bilevel Nonsmooth Optimization

被引:0
|
作者
T., Okuno
A., Takeda
A., Kawana
M., Watanabe
机构
[1] Center for Advanced Intelligence Project, RIKEN, Tokyo,103-0027, Japan
[2] Graduate School of Information Science and Technology, The University of Tokyo, Tokyo,113-8656, Japan
[3] Center for Advanced Intelligence Project, RIKEN, Tokyo,103-0027, Japan
[4] Department of Industrial Engineering and Economics, Tokyo Institute of Technology, Tokyo,152-8550, Japan
[5] Department of Mathematical Informatics, The University of Tokyo, Tokyo,113-8656, Japan
基金
日本学术振兴会;
关键词
Computational methods;
D O I
暂无
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We propose a bilevel optimization strategy for selecting the best hyperparameter value for the nonsmooth `p regularizer with 0 p-regularized problem as the lower-level problem. Despite the recent popularity of nonconvex `p-regularizer and the usefulness of bilevel optimization for selecting hyperparameters, algorithms for such bilevel problems have not been studied because of the difficulty of `p-regularizer. Our contribution is the proposal of the first algorithm equipped with a theoretical guarantee for finding the best hyperparameter of `p-regularized supervised learning problems. Specifically, we propose a smoothing-type algorithm for the above mentioned bilevel optimization problems and provide a theoretical convergence guarantee for the algorithm. Indeed, since optimality conditions are not known for such bilevel optimization problems so far, new necessary optimality conditions, which are called the SB-KKT conditions, are derived and it is shown that a sequence generated by the proposed algorithm actually accumulates at a point satisfying the SB-KKT conditions under some mild assumptions. The proposed algorithm is simple and scalable as our numerical comparison to Bayesian optimization and grid search indicates. ©2021 Takayuki Okuno, Akiko Takeda, Akihiro Kawana, and Motokazu Watanabe.
引用
收藏
页码:1 / 47
相关论文
共 50 条
  • [41] Hyperparameter optimization via sequential uniform designs
    Yang, Zebin
    Zhang, Aijun
    Journal of Machine Learning Research, 2021, 22
  • [42] Stabilization via nonsmooth, nonconvex optimization
    Burke, James V.
    Henrion, Didier
    Lewis, Adrian S.
    Overton, Michael L.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2006, 51 (11) : 1760 - 1769
  • [43] Certificates of infeasibility via nonsmooth optimization
    Fendl, Hannes
    Neumaier, Arnold
    Schichl, Hermann
    JOURNAL OF GLOBAL OPTIMIZATION, 2017, 69 (01) : 157 - 182
  • [44] Certificates of infeasibility via nonsmooth optimization
    Hannes Fendl
    Arnold Neumaier
    Hermann Schichl
    Journal of Global Optimization, 2017, 69 : 157 - 182
  • [45] Machine Learning Assisted Hyperparameter Tuning for Optimization
    Linkous, Lauren
    Lundquist, Jonathan
    Suche, Michael
    Topsakal, Erdem
    2024 IEEE INC-USNC-URSI RADIO SCIENCE MEETING (JOINT WITH AP-S SYMPOSIUM), 2024, : 107 - 108
  • [46] Reinforcement Learning for Model Selection and Hyperparameter Optimization
    Wu J.
    Chen S.-P.
    Chen X.-Y.
    Zhou R.
    Dianzi Keji Daxue Xuebao/Journal of the University of Electronic Science and Technology of China, 2020, 49 (02): : 255 - 261
  • [47] Meta Learning for Hyperparameter Optimization in Dialogue System
    Chien, Jen-Tzung
    Lieow, Wei Xiang
    INTERSPEECH 2019, 2019, : 839 - 843
  • [48] Sherpa: Robust hyperparameter optimization for machine learning
    Hertel, Lars
    Collado, Julian
    Sadowski, Peter
    Ott, Jordan
    Baldi, Pierre
    SOFTWAREX, 2020, 12
  • [49] Hyperparameter optimization for machine learning models based on Bayesian optimization
    Wu J.
    Chen X.-Y.
    Zhang H.
    Xiong L.-D.
    Lei H.
    Deng S.-H.
    Journal of Electronic Science and Technology, 2019, 17 (01) : 26 - 40
  • [50] Hyperparameter Optimization for Machine Learning Models Based on Bayesian Optimization
    Jia Wu
    Xiu-Yun Chen
    Hao Zhang
    Li-Dong Xiong
    Hang Lei
    Si-Hao Deng
    Journal of Electronic Science and Technology, 2019, (01) : 26 - 40