HAMLET - A Learning Curve-Enabled Multi-Armed Bandit for Algorithm Selection

被引:1
|
作者
Schmidt, Mischa [1 ]
Gastinger, Julia [1 ]
Nicolas, Sebastien [1 ]
Schuelke, Anett [1 ]
机构
[1] NEC Labs Europe GmbH, Kurfursten Anlage 36, D-69115 Heidelberg, Germany
关键词
Automated Machine Learning; Multi-Armed Bandit; Learning Curve Extrapolation;
D O I
10.1109/ijcnn48605.2020.9207233
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Automated algorithm selection and hyperparameter tuning facilitates the application of machine learning. Traditional multi-armed bandit strategies look to the history of observed rewards to identify the most promising arms for optimizing expected total reward in the long run. When considering limited time budgets and computational resources, this backward view of rewards is inappropriate as the bandit should look into the future for anticipating the highest final reward at the end of a specified time budget. This work addresses that insight by introducing HAMLET, which extends the bandit approach with learning curve extrapolation and computation time-awareness for selecting among a set of machine learning algorithms. Results show that the HAMLET Variants 1-3 exhibit equal or better performance than other bandit-based algorithm selection strategies in experiments with recorded hyperparameter tuning traces for the majority of considered time budgets. The best performing HAMLET Variant 3 combines learning curve extrapolation with the well-known upper confidence bound exploration bonus. That variant performs better than all non-HAMLET policies with statistical significance at the 95% level for 1,485 runs.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] A Multi-Armed Bandit Strategy for Countermeasure Selection
    Cochrane, Madeleine
    Hunjet, Robert
    [J]. 2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 2510 - 2515
  • [2] DBA: Dynamic Multi-Armed Bandit Algorithm
    Nobari, Sadegh
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9869 - 9870
  • [3] Learning State Selection for Reconfigurable Antennas: A Multi-Armed Bandit Approach
    Gulati, Nikhil
    Dandekar, Kapil R.
    [J]. IEEE TRANSACTIONS ON ANTENNAS AND PROPAGATION, 2014, 62 (03) : 1027 - 1038
  • [4] CONTEXTUAL MULTI-ARMED BANDIT ALGORITHMS FOR PERSONALIZED LEARNING ACTION SELECTION
    Manickam, Indu
    Lan, Andrew S.
    Baraniuk, Richard G.
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 6344 - 6348
  • [5] Automated Collaborator Selection for Federated Learning with Multi-armed Bandit Agents
    Larsson, Hannes
    Riaz, Hassam
    Ickin, Selim
    [J]. PROCEEDINGS OF THE 4TH FLEXNETS WORKSHOP ON FLEXIBLE NETWORKS, ARTIFICIAL INTELLIGENCE SUPPORTED NETWORK FLEXIBILITY AND AGILITY (FLEXNETS'21), 2021, : 44 - 49
  • [6] Multi-armed Bandit Algorithm against Strategic Replication
    Shin, Suho
    Lee, Seungjoon
    Ok, Jungseul
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151 : 403 - 431
  • [7] The multi-armed bandit, with constraints
    Eric V. Denardo
    Eugene A. Feinberg
    Uriel G. Rothblum
    [J]. Annals of Operations Research, 2013, 208 : 37 - 62
  • [8] The Assistive Multi-Armed Bandit
    Chan, Lawrence
    Hadfield-Menell, Dylan
    Srinivasa, Siddhartha
    Dragan, Anca
    [J]. HRI '19: 2019 14TH ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2019, : 354 - 363
  • [9] The multi-armed bandit, with constraints
    Denardo, Eric V.
    Feinberg, Eugene A.
    Rothblum, Uriel G.
    [J]. ANNALS OF OPERATIONS RESEARCH, 2013, 208 (01) : 37 - 62
  • [10] Multi-armed bandit games
    Gursoy, Kemal
    [J]. ANNALS OF OPERATIONS RESEARCH, 2024,