Adaptive Active Learning as a Multi-armed Bandit Problem

被引:2
|
作者
Czarnecki, Wojciech M. [1 ]
Podolak, Igor T. [1 ]
机构
[1] Jagiellonian Univ, Fac Math & Comp Sci, Krakow, Poland
来源
21ST EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE (ECAI 2014) | 2014年 / 263卷
关键词
D O I
10.3233/978-1-61499-419-0-989
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we present a new active learning strategy whose main focus is to have the ability to adapt to the unknown (or changing) learning scenario. We introduce the learners' ensemble based approach and model it as the multi-armed bandit problem. Presented application of simple exploration-exploitation trade-off algorithms from the UCB and EXP3 families show an improvement over using the classical strategies. Evaluation on data from UCI database compare three different selection algorithms. In our tests, presented method shows promising results.
引用
收藏
页码:989 / 990
页数:2
相关论文
共 50 条
  • [11] Adaptive Algorithm for Multi-Armed Bandit Problem with High-Dimensional Covariates
    Qian, Wei
    Ing, Ching-Kang
    Liu, Ji
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2024, 119 (546) : 970 - 982
  • [12] Achieving Fairness in the Stochastic Multi-Armed Bandit Problem
    Patil, Vishakha
    Ghalme, Ganesh
    Nair, Vineet
    Narahari, Y.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [13] On the Combinatorial Multi-Armed Bandit Problem with Markovian Rewards
    Gai, Yi
    Krishnamachari, Bhaskar
    Liu, Mingyan
    2011 IEEE GLOBAL TELECOMMUNICATIONS CONFERENCE (GLOBECOM 2011), 2011,
  • [14] Possibilistic reward methods for the multi-armed bandit problem
    Martin, Miguel
    Jimenez-Martin, Antonio
    Mateos, Alfonso
    NEUROCOMPUTING, 2018, 310 : 201 - 212
  • [15] Achieving Fairness in the Stochastic Multi-Armed Bandit Problem
    Patil, Vishakha
    Ghalme, Ganesh
    Nair, Vineet
    Narahari, Y.
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 5379 - 5386
  • [16] The sample complexity of exploration in the multi-armed bandit problem
    Mannor, S
    Tsitsiklis, JN
    JOURNAL OF MACHINE LEARNING RESEARCH, 2004, 5 : 623 - 648
  • [17] Interface Design Optimization as a Multi-Armed Bandit Problem
    Lomas, J. Derek
    Forlizzi, Jodi
    Poonwala, Nikhil
    Patel, Nirmal
    Shodhan, Sharan
    Patel, Kishan
    Koedinger, Ken
    Brunskill, Emma
    34TH ANNUAL CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2016, 2016, : 4142 - 4153
  • [18] Scalable Discrete Sampling as a Multi-Armed Bandit Problem
    Chen, Yutian
    Ghahramani, Zoubin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [19] Online Optimization Algorithms for Multi-Armed Bandit Problem
    Kamalov, Mikhail
    Dobrynin, Vladimir
    Balykina, Yulia
    2017 CONSTRUCTIVE NONSMOOTH ANALYSIS AND RELATED TOPICS (DEDICATED TO THE MEMORY OF V.F. DEMYANOV) (CNSA), 2017, : 141 - 143
  • [20] THE MULTI-ARMED BANDIT PROBLEM: AN EFFICIENT NONPARAMETRIC SOLUTION
    Chan, Hock Peng
    ANNALS OF STATISTICS, 2020, 48 (01): : 346 - 373