Online Kernel Selection via Grouped Adversarial Bandit Model

被引:0
|
作者
Li, Junfan [1 ]
Liao, Shizhong [1 ]
机构
[1] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300350, Peoples R China
基金
中国国家自然科学基金;
关键词
online kernel selection; sequential decision; group bandit model; regret; time complexity; REGRET;
D O I
10.1109/ICTAI.2019.00100
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study kernel selection for online kernel learning, also known as online kernel selection which can be treated as a sequential decision problem and thus must balance the regret and the time complexity. Existing online kernel selection approaches via expert advice and classical adversarial bandit model can not meet the issue. In this work, we propose a novel grouped adversarial bandit solution to the problem. We first correspond each candidate kernel to a basic arm of an adversarial bandit problem. Then, all of the kernels are divided into several groups where each group is abstracted as a super arm. At each round, we choose a super arm and a basic kernel within the selected super arm, and make prediction by an online kernel learning algorithm. Besides, we introduce a Bernoulli random variable to decide whether to choose all of the rest super arms. Theoretical analysis shows the proposed approach balances the regret and the time complexity explicitly, which could enjoy better pseudo-regret and high probability regret bound than classical adversarial bandit model and lighter time complexity than expert advice model. Experimental results on benchmark datasets verify that the proposed approach balances the efficiency and effectiveness better.
引用
收藏
页码:682 / 689
页数:8
相关论文
共 50 条
  • [31] A Kernel Multiple Change-point Algorithm via Model Selection
    Arlot, Sylvain
    Celisse, Alain
    Harchaoui, Zaid
    JOURNAL OF MACHINE LEARNING RESEARCH, 2019, 20
  • [32] Stepwise Model Selection for Sequence Prediction via Deep Kernel Learning
    Zhang, Yao
    Jarrett, Daniel
    van der Schaar, Mihaela
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 2304 - 2313
  • [33] Online Evaluation of Audiences for Targeted Advertising via Bandit Experiments
    Geng, Tong
    Lin, Xiliang
    Nair, Harikesh S.
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 13273 - 13279
  • [34] ADVERSARIAL BANDIT FOR ONLINE INTERACTIVE ACTIVE LEARNING OF ZERO-SHOT SPOKEN LANGUAGE UNDERSTANDING
    Ferreira, Emmanuel
    Masson, Alexandre Reiffers
    Jabaian, Bassam
    Lefevre, Fabrice
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 6155 - 6159
  • [35] Model Selection for Production System via Automated Online Experiments
    Dai, Zhenwen
    Ravichandran, Praveen
    Fazelnia, Ghazal
    Carterette, Ben
    Lalmas-Roelleke, Mounia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [36] Parameter-Free Online Learning via Model Selection
    Foster, Dylan J.
    Kale, Satyen
    Mohri, Mehryar
    Sridharan, Karthik
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [38] Graph-based Adversarial Online Kernel Learning with Adaptive Embedding
    Yang, Peng
    Li, Xiaoyun
    Li, Ping
    2021 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2021), 2021, : 797 - 806
  • [39] A neural model for visual selection of grouped spatial arrays
    Domijan, D
    NEUROREPORT, 2003, 14 (03) : 367 - 370
  • [40] On the grouped selection and model complexity of the adaptive elastic net
    Ghosh, Samiran
    STATISTICS AND COMPUTING, 2011, 21 (03) : 451 - 462