Meta Learning for Hyperparameter Optimization in Dialogue System

被引:22
|
作者
Chien, Jen-Tzung [1 ]
Lieow, Wei Xiang [1 ]
机构
[1] Natl Chiao Tung Univ, Dept Elect & Comp Engn, Hsinchu, Taiwan
来源
关键词
dialogue system; meta learning; Bayesian optimization; recurrent neural network;
D O I
10.21437/Interspeech.2019-1383
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
The performance of dialogue system based on deep reinforcement learning (DRL) highly depends on the selected hyperparameters in DRL algorithms. Traditionally, Gaussian process (GP) provides a probabilistic approach to Bayesian optimization for sequential search which is beneficial to select optimal hyperparameter. However, GP suffers from the expanding computation when the dimension of hyperparameters and the number of search points are increased. This paper presents a meta learning approach to carry out multifidelity Bayesian optimization where a two-level recurrent neural network (RNN) is developed for sequential learning and optimization. The search space is explored via the first-level RNN with cheap and low fidelity over a global region of hyperparameters. The optimization is then exploited and leveraged by the second-level RNN with a high fidelity on the successively small regions. The experiments on the hyperparameter optimization for dialogue system based on the deep Q network show the effectiveness and efficiency by using the proposed multifidelity Bayesian optimization.
引用
收藏
页码:839 / 843
页数:5
相关论文
共 50 条
  • [1] Bilevel Programming for Hyperparameter Optimization and Meta-Learning
    Franceschi, Luca
    Frasconi, Paolo
    Salzo, Saverio
    Grazzi, Riccardo
    Pontil, Massimilano
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [2] Initializing Bayesian Hyperparameter Optimization via Meta-Learning
    Feurer, Matthias
    Springenberg, Jost Tobias
    Hutter, Frank
    PROCEEDINGS OF THE TWENTY-NINTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2015, : 1128 - 1135
  • [3] Ensemble Clustering based on Meta-Learning and Hyperparameter Optimization
    Treder-Tschechlov, Dennis
    Fritz, Manuel
    Schwarz, Holger
    Mitschang, Bernhard
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2024, 17 (11): : 2880 - 2892
  • [4] Model-Based Meta-reinforcement Learning for Hyperparameter Optimization
    Albrechts, Jeroen
    Martin, Hugo M.
    Tavakol, Maryam
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2024, PT I, 2025, 15346 : 27 - 39
  • [5] Hyperparameter optimization in learning systems
    Andonie, Razvan
    JOURNAL OF MEMBRANE COMPUTING, 2019, 1 (04) : 279 - 291
  • [6] Learning Hyperparameter Optimization Initializations
    Wistuba, Martin
    Schilling, Nicolas
    Schmidt-Thieme, Lars
    PROCEEDINGS OF THE 2015 IEEE INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (IEEE DSAA 2015), 2015, : 339 - 348
  • [7] Hyperparameter optimization in learning systems
    Răzvan Andonie
    Journal of Membrane Computing, 2019, 1 : 279 - 291
  • [8] Hyperparameter Optimization for Improving Recognition Efficiency of an Adaptive Learning System
    Tran, Diem-Phuc
    Nguyen, Gia-Nhu
    Hoang, Van-Dung
    IEEE ACCESS, 2020, 8 : 160569 - 160580
  • [9] EvoGrad: Efficient Gradient-Based Meta-Learning and Hyperparameter Optimization
    Bohdal, Ondrej
    Yang, Yongxin
    Hospedales, Timothy
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [10] Meta-Surrogate Benchmarking for Hyperparameter Optimization
    Klein, Aaron
    Dai, Zhenwen
    Hutter, Frank
    Lawrence, Neil
    Gonzalez, Javier
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32