Bayesian Optimization Meets Bayesian Optimal Stopping

被引:0
|
作者
Dai, Zhongxiang [1 ]
Yu, Haibin [1 ]
Low, Bryan Kian Hsiang [1 ]
Jaillet, Patrick [2 ]
机构
[1] Natl Univ Singapore, Dept Comp Sci, Singapore, Singapore
[2] MIT, Dept Elect Engn & Comp Sci, Cambridge, MA 02139 USA
基金
新加坡国家研究基金会;
关键词
DECENTRALIZED DATA FUSION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Bayesian optimization (BO) is a popular paradigm for optimizing the hyperparameters of machine learning (ML) models due to its sample efficiency. Many ML models require running an iterative training procedure (e.g., stochastic gradient descent). This motivates the question whether information available during the training process (e.g., validation accuracy after each epoch) can be exploited for improving the epoch efficiency of BO algorithms by early-stopping model training under hyperparameter settings that will end up under-performing and hence eliminating unnecessary training epochs. This paper proposes to unify BO (specifically, Gaussian process-upper confidence bound (GP-UCB)) with Bayesian optimal stopping (BO-BOS) to boost the epoch efficiency of BO. To achieve this, while GP-UCB is sample-efficient in the number of function evaluations, BOS complements it with epoch efficiency for each function evaluation by providing a principled optimal stopping mechanism for early stopping. BO-BOS preserves the (asymptotic) no-regret performance of GP-UCB using our specified choice of BOS parameters that is amenable to an elegant interpretation in terms of the exploration-exploitation trade-off. We empirically evaluate the performance of BO-BOS and demonstrate its generality in hyperparameter optimization of ML models and two other interesting applications.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Bayesian stopping
    Douven, Igor
    [J]. JOURNAL OF MATHEMATICAL PSYCHOLOGY, 2023, 116
  • [2] Bayesian Optimization Meets Self-Distillation
    Lee, HyunJae
    Song, Heon
    Lee, Hyeonsoo
    Lee, Gi-hyeon
    Park, Suyeong
    Yoo, Donggeun
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 1696 - 1705
  • [3] Asymptotically pointwise optimal and asymptotically optimal stopping times in the Bayesian inference
    Alicja Jokiel-Rokita
    [J]. Statistical Papers, 2008, 49 : 165 - 175
  • [4] Asymptotically pointwise optimal and asymptotically optimal stopping times in the Bayesian inference
    Jokiel-Rokita, Alicja
    [J]. STATISTICAL PAPERS, 2008, 49 (02) : 165 - 175
  • [5] BAYESIAN STOPPING RULES FOR MULTISTART GLOBAL OPTIMIZATION METHODS
    BOENDER, CGE
    KAN, AHGR
    [J]. MATHEMATICAL PROGRAMMING, 1987, 37 (01) : 59 - 80
  • [6] Bayesian Optimization Meets Riemannian Manifolds in Robot Learning
    Jaquier, Noemie
    Rozo, Leonel
    Calinon, Sylvain
    Burger, Mathias
    [J]. CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [7] Bayesian Optimization of MOSFET Devices Using Effective Stopping Condition
    Kim, Bokyeom
    Shin, Mincheol
    [J]. IEEE ACCESS, 2021, 9 : 108480 - 108494
  • [8] Automating Bayesian optimization with Bayesian optimization
    Malkomes, Gustavo
    Garnett, Roman
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [9] PAC-Bayesian Theory Meets Bayesian Inference
    Germain, Pascal
    Bach, Francis
    Lacoste, Alexandre
    Lacoste-Julien, Simon
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [10] Bayesian stopping rules for trials
    Nicholl, J
    Goodacre, S
    [J]. LANCET, 2002, 359 (9300): : 76 - 76