Quantum greedy algorithms for multi-armed bandits

被引:0
|
作者
Ohno, Hiroshi [1 ]
机构
[1] Toyota Cent Res & Dev Labs Inc, 41-1 Yokomichi, Nagakute, Aichi 4801192, Japan
关键词
Multi-armed bandits; -greedy algorithm; MovieLens dataset; Quantum amplitude amplification; Regret analysis;
D O I
10.1007/s11128-023-03844-2
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Multi-armed bandits are widely used in machine learning applications such as recommendation systems. Here, we implement two quantum versions of the e-greedy algorithm, a popular algorithm for multi-armed bandits. One of the quantum greedy algorithms uses a quantum maximization algorithm and the other is a simple algorithm that uses an amplitude encoding method as a quantum subroutine instead of the argmax operation in the e-greedy algorithm. For the former algorithm, given a quantum oracle, the query complexity is on the order root K (O(root K)) in each round, where K is the number of arms. For the latter algorithm, quantum parallelism is achieved by the quantum superposition of the arms and the run-time complexity is on the order O(K)/O(log K) in each round. Bernoulli reward distributions and the MovieLens dataset are used to evaluate the algorithms with their classical counterparts. The experimental results show that for best arm identification, the performance of the quantum greedy algorithm is comparable with that of the classical counterparts.
引用
收藏
页数:20
相关论文
共 50 条
  • [21] Quantum Multi-Armed Bandits and Stochastic Linear Bandits Enjoy Logarithmic Regrets
    Wan, Zongqi
    Zhang, Zhijie
    Li, Tongyang
    Zhang, Jialin
    Sun, Xiaoming
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 10087 - 10094
  • [22] Aggregation of Multi-Armed Bandits Learning Algorithms for Opportunistic Spectrum Access
    Besson, Lilian
    Kaufmann, Emilie
    Moy, Christophe
    2018 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2018,
  • [23] Multi-armed bandits for performance marketing
    Gigli, Marco
    Stella, Fabio
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2024,
  • [24] Lenient Regret for Multi-Armed Bandits
    Merlis, Nadav
    Mannor, Shie
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8950 - 8957
  • [25] Finding structure in multi-armed bandits
    Schulz, Eric
    Franklin, Nicholas T.
    Gershman, Samuel J.
    COGNITIVE PSYCHOLOGY, 2020, 119
  • [26] ON MULTI-ARMED BANDITS AND DEBT COLLECTION
    Czekaj, Lukasz
    Biegus, Tomasz
    Kitlowski, Robert
    Tomasik, Pawel
    36TH ANNUAL EUROPEAN SIMULATION AND MODELLING CONFERENCE, ESM 2022, 2022, : 137 - 141
  • [27] Visualizations for interrogations of multi-armed bandits
    Keaton, Timothy J.
    Sabbaghi, Arman
    STAT, 2019, 8 (01):
  • [28] Multi-armed bandits with dependent arms
    Singh, Rahul
    Liu, Fang
    Sun, Yin
    Shroff, Ness
    MACHINE LEARNING, 2024, 113 (01) : 45 - 71
  • [29] On Kernelized Multi-Armed Bandits with Constraints
    Zhou, Xingyu
    Ji, Bo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [30] Multi-Armed Bandits in Metric Spaces
    Kleinberg, Robert
    Slivkins, Aleksandrs
    Upfal, Eli
    STOC'08: PROCEEDINGS OF THE 2008 ACM INTERNATIONAL SYMPOSIUM ON THEORY OF COMPUTING, 2008, : 681 - +