Gorthaur : A Portfolio Approach for Dynamic Selection of Multi-Armed Bandit Algorithms for Recommendation

被引:5
|
作者
Gutowski, Nicolas [1 ,2 ]
Amghar, Tassadit [1 ]
Camp, Olivier [2 ]
Chhel, Fabien [2 ]
机构
[1] Univ Angers, LERIA, Angers, France
[2] ESEO TECH, Angers, France
关键词
Application of Reinforcement Learning; Contextual Multi-Armed Bandit; Recommender Systems; Portfolio Approach; Heuristic; CONTEXT;
D O I
10.1109/ICTAI.2019.00161
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recommendation systems must reach a good global accuracy but also diversify their recommendations. Despite the theoretically grounded guarantees, we observe that multi-armed bandit algorithms obtain different results depending on the nature of the real-world applications or the offline datasets that is used. Thus, before choosing an algorithm, it is necessary to carry out a preliminary offline evaluation on the criteria of global accuracy and, if necessary, diversity. However, recommendation systems are notoriously hard to evaluate due to their interactive and dynamic nature. Hence, we have implemented a portfolio approach, entitled Gorthaur, which uses a heuristic to dynamically select multi-armed bandits algorithms used for recommending. Thus, Gorthaur aims at selecting algorithms by maximising the two criteria of global accuracy and diversity. Following our results, we observe that the advantage of using Gorthaur is twofold: 1) Find a trade-off in cases where there is no prior knowledge about the nature of the dataset or the recommendation application we want to deploy; 2) Rapidly shed light on a set of optimal algorithms.
引用
收藏
页码:1164 / 1171
页数:8
相关论文
共 50 条
  • [1] Scaling Multi-Armed Bandit Algorithms
    Fouche, Edouard
    Komiyama, Junpei
    Boehm, Klemens
    [J]. KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, : 1449 - 1459
  • [2] Operator Selection using Improved Dynamic Multi-Armed Bandit
    Belluz, Jany
    Gaudesi, Marco
    Squillero, Giovanni
    Tonda, Alberto
    [J]. GECCO'15: PROCEEDINGS OF THE 2015 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2015, : 1311 - 1317
  • [3] Dynamic Multi-Armed Bandit with Covariates
    Pavlidis, Nicos G.
    Tasoulis, Dimitris K.
    Adams, Niall M.
    Hand, David J.
    [J]. ECAI 2008, PROCEEDINGS, 2008, 178 : 777 - +
  • [4] Multi-armed bandit algorithms and empirical evaluation
    Vermorel, J
    Mohri, M
    [J]. MACHINE LEARNING: ECML 2005, PROCEEDINGS, 2005, 3720 : 437 - 448
  • [5] CONTEXTUAL MULTI-ARMED BANDIT ALGORITHMS FOR PERSONALIZED LEARNING ACTION SELECTION
    Manickam, Indu
    Lan, Andrew S.
    Baraniuk, Richard G.
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 6344 - 6348
  • [6] Risk-aware multi-armed bandit problem with application to portfolio selection
    Huo, Xiaoguang
    Fu, Feng
    [J]. ROYAL SOCIETY OPEN SCIENCE, 2017, 4 (11):
  • [7] Anytime Algorithms for Multi-Armed Bandit Problems
    Kleinberg, Robert
    [J]. PROCEEDINGS OF THE SEVENTHEENTH ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, 2006, : 928 - 936
  • [8] A Multi-Armed Bandit Model Selection for Cold-Start User Recommendation
    Felicio, Cricia Z.
    Paixao, Klerisson V. R.
    Barcelos, Celia A. Z.
    Preux, Philippe
    [J]. PROCEEDINGS OF THE 25TH CONFERENCE ON USER MODELING, ADAPTATION AND PERSONALIZATION (UMAP'17), 2017, : 32 - 40
  • [9] Contextual Multi-Armed Bandit for Email Layout Recommendation
    Chen, Yan
    Vankov, Emilian
    Baltrunas, Linas
    Donovan, Preston
    Mehta, Akash
    Schroeder, Benjamin
    Herman, Matthew
    [J]. PROCEEDINGS OF THE 17TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2023, 2023, : 400 - 402
  • [10] Dynamic clustering based contextual combinatorial multi-armed bandit for online recommendation
    Yan, Cairong
    Han, Haixia
    Zhang, Yanting
    Zhu, Dandan
    Wan, Yongquan
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 257