Scalable Bayesian Optimization Using Deep Neural Networks

被引:0
|
作者
Snoek, Jasper [1 ]
Rippel, Oren [1 ,2 ]
Swersky, Kevin [3 ]
Kiros, Ryan [3 ]
Satish, Nadathur [4 ]
Sundaram, Narayanan [4 ]
Patwary, Md. Mostofa Ali [4 ]
Prabhat [5 ]
Adams, Ryan P. [1 ]
机构
[1] Harvard Univ, Sch Engn & Appl Sci, Cambridge, MA 02138 USA
[2] MIT, Dept Math, Cambridge, MA 02139 USA
[3] Univ Toronto, Dept Comp Sci, Toronto, ON, Canada
[4] Intel Labs, Parallel Comp Lab, Santa Clara, CA USA
[5] Lawrence Berkeley Natl Lab, NERSC, Berkeley, CA 94720 USA
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Bayesian optimization is an effective methodology for the global optimization of functions with expensive evaluations. It relies on querying a distribution over functions defined by a relatively cheap surrogate model. An accurate model for this distribution over functions is critical to the effectiveness of the approach, and is typically fit using Gaussian processes (GPs). However, since GPs scale cubically with the number of observations, it has been challenging to handle objectives whose optimization requires many evaluations, and as such, massively parallelizing the optimization. In this work, we explore the use of neural networks as an alternative to GPs to model distributions over functions. We show that performing adaptive basis function regression with a neural network as the parametric form performs competitively with state-of-the-art GP-based approaches, but scales linearly with the number of data rather than cubically. This allows us to achieve a previously intractable degree of parallelism, which we apply to large scale hyperparameter optimization, rapidly finding competitive models on benchmark object recognition tasks using convolutional networks, and image caption generation using neural language models.
引用
收藏
页码:2171 / 2180
页数:10
相关论文
共 50 条
  • [1] Efficient Priors for Scalable Variational Inference in Bayesian Deep Neural Networks
    Krishnan, Ranganath
    Subedar, Mahesh
    Tickoo, Omesh
    Labs, Intel
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 773 - 777
  • [2] Heterogeneous gradient computing optimization for scalable deep neural networks
    Moreno-Alvarez, Sergio
    Paoletti, Mercedes E.
    Rico-Gallego, Juan A.
    Haut, Juan M.
    [J]. JOURNAL OF SUPERCOMPUTING, 2022, 78 (11): : 13455 - 13469
  • [3] Heterogeneous gradient computing optimization for scalable deep neural networks
    Sergio Moreno-Álvarez
    Mercedes E. Paoletti
    Juan A. Rico-Gallego
    Juan M. Haut
    [J]. The Journal of Supercomputing, 2022, 78 : 13455 - 13469
  • [4] Scalable Object Detection using Deep Neural Networks
    Erhan, Dumitru
    Szegedy, Christian
    Toshev, Alexander
    Anguelov, Dragomir
    [J]. 2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, : 2155 - 2162
  • [5] Scalable Gaussian Process Regression Using Deep Neural Networks
    Huang, Wenbing
    Zhao, Deli
    Sun, Fuchun
    Liu, Huaping
    Chang, Edward
    [J]. PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 3576 - 3582
  • [6] A scalable model of vegetation transitions using deep neural networks
    Rammer, Werner
    Seidl, Rupert
    [J]. METHODS IN ECOLOGY AND EVOLUTION, 2019, 10 (06): : 879 - 890
  • [7] Development of Deep Residual Neural Networks for Gear Pitting Fault Diagnosis Using Bayesian Optimization
    Li, Jialin
    Chen, Renxiang
    Huang, Xianzhen
    Qu, Yongzhi
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71
  • [8] Modeling Bitcoin Prices using Signal Processing Methods, Bayesian Optimization, and Deep Neural Networks
    Bhaskar Tripathi
    Rakesh Kumar Sharma
    [J]. Computational Economics, 2023, 62 : 1919 - 1945
  • [9] Modeling Bitcoin Prices using Signal Processing Methods, Bayesian Optimization, and Deep Neural Networks
    Tripathi, Bhaskar
    Sharma, Rakesh Kumar
    [J]. COMPUTATIONAL ECONOMICS, 2023, 62 (04) : 1919 - 1945
  • [10] Basic Enhancement Strategies When Using Bayesian Optimization for Hyperparameter Tuning of Deep Neural Networks
    Cho, Hyunghun
    Kim, Yongjin
    Lee, Eunjung
    Choi, Daeyoung
    Lee, Yongjae
    Rhee, Wonjong
    [J]. IEEE ACCESS, 2020, 8 : 52588 - 52608