Hyperparameter Transfer Learning with Adaptive Complexity

被引:0
|
作者
Horvath, Samuel [1 ]
Klein, Aaron [2 ]
Richtarik, Peter [1 ]
Archambeau, Cedric [2 ]
机构
[1] KAUST, Thuwal, Saudi Arabia
[2] Amazon Web Serv, Seattle, WA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Bayesian optimization (BO) is a sample efficient approach to automatically tune the hyperparameters of machine learning models. In practice, one frequently has to solve similar hyperparameter tuning problems sequentially. For example, one might have to tune a type of neural network learned across a series of different classification problems. Recent work on multi-task BO exploits knowledge gained from previous tuning tasks to speed up a new tuning task. However, previous approaches do not account for the fact that BO is a sequential decision making procedure. Hence, there is in general a mismatch between the number of evaluations collected in the current tuning task compared to the number of evaluations accumulated in all previously completed tasks. In this work, we enable multi-task BO to compensate for this mismatch, such that the transfer learning procedure is able to handle different data regimes in a principled way. We propose a new multi-task BO method that learns a set of ordered, non-linear basis functions of increasing complexity via nested drop-out and automatic relevance determination. Experiments on a variety of hyperparameter tuning problems show that our method improves the sample efficiency of recently published multi-task BO methods.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Scalable Hyperparameter Transfer Learning
    Perrone, Valerio
    Jenatton, Rodolphe
    Seeger, Matthias
    Archambeau, Cedric
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [2] Hyperparameter Learning via Distributional Transfer
    Law, Ho Chung Leon
    Zhao, Peilin
    Chan, Lucian
    Huang, Junzhou
    Sejdinovic, Dino
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [3] Hyperparameter Search for Machine Learning Algorithms for Optimizing the Computational Complexity
    Ali, Yasser A.
    Awwad, Emad Mahrous
    Al-Razgan, Muna
    Maarouf, Ali
    [J]. PROCESSES, 2023, 11 (02)
  • [4] Efficient Transfer Learning Method for Automatic Hyperparameter Tuning
    Yogatama, Dani
    Mann, Gideon
    [J]. ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 33, 2014, 33 : 1077 - 1085
  • [5] Deep hyperparameter transfer learning for diabetic retinopathy classification
    Patil, Mahesh S.
    Chickerur, Satyadhyan
    Kumar, Yeshwanth V. S.
    Bakale, Vijayalakshmi A.
    Giraddi, Shantala
    Roodagi, Vivekanand C.
    Kulkarni, Yashaswini N.
    [J]. TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2021, 29 : 2824 - 2839
  • [6] Hyperparameter Learning for Conditional Kernel Mean Embeddings with Rademacher Complexity Bounds
    Hsu, Kelvin
    Nock, Richard
    Ramos, Fabio
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2018, PT II, 2019, 11052 : 227 - 242
  • [7] Hyperparameter Optimization for Improving Recognition Efficiency of an Adaptive Learning System
    Tran, Diem-Phuc
    Nguyen, Gia-Nhu
    Hoang, Van-Dung
    [J]. IEEE ACCESS, 2020, 8 : 160569 - 160580
  • [8] Adaptive Learning: Context and Complexity
    Dziuban, Charles
    Howlin, Colm
    Moskal, Patsy
    Johnson, Connie
    Eid, Mitchell
    Kmetz, Brandon
    [J]. E-MENTOR, 2018, (05): : 13 - 23
  • [9] An Adaptive Hyperparameter Strategy Optimization Method for Spacecraft Rendezvous and Orbital Transfer
    Sun, Leixiang
    Guo, Yanning
    Deng, Wudong
    Lyu, Yueyong
    Ma, Guangfu
    [J]. Yuhang Xuebao/Journal of Astronautics, 2024, 45 (01): : 52 - 62
  • [10] Learning search spaces for Bayesian optimization: Another view of hyperparameter transfer learning
    Perrone, Valerio
    Shen, Huibin
    Seeger, Matthias
    Archambeau, Cedric
    Jenatton, Rodolphe
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32