Enhancing Noisy Binary Search Efficiency through Deep Reinforcement Learning

被引:0
|
作者
Ma, Rui [1 ]
Tao, Yudong [1 ]
Khodeiry, Mohamed M. [2 ]
Alawa, Karam A. [2 ]
Shyu, Mei-Ling [3 ]
Lee, Richard K. [1 ,2 ]
机构
[1] Univ Miami, Dept Elect & Comp Engn, Coral Gables, FL USA
[2] Univ Miami, Bascom Palmer Eye Inst, Miller Sch Med, Miami, FL USA
[3] Univ Missouri Kansas City, Sch Sci & Engn, Kansas City, MO USA
关键词
Deep learning; deep reinforcement learning; noisy binary search;
D O I
10.1109/IRI58017.2023.00033
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Noisy binary search (NBS) aims to find the closest element to a target value within a sorted array through erroneous queries. In an ideal NBS environment where the error rate remains constant, and the costs of all queries are the same, the maximum likelihood estimation (MLE) procedure has been proven to be the optimal decision strategy. However, in some non-ideal NBS problems, both the error rates and the costs are dependent on the queries, and in some cases, finding the optimal decision strategies can be intractable. We propose to use deep reinforcement learning to approximate the optimal decision strategy in the NBS problem, in which an intelligent agent is used to interact with the NBS environment. A dueling double deep Q-network guides the agent to take action at each step, either to generate a query or to stop the search and predict the target value. An optimized policy will be derived by training the network in the NBS environment until convergence. By evaluating our proposed algorithm on a non-ideal NBS environment, visual field test, we show that the performance of our proposed algorithm surpasses baseline visual field testing algorithms by a large margin.
引用
收藏
页码:154 / 159
页数:6
相关论文
共 50 条
  • [1] Efficient Novelty Search Through Deep Reinforcement Learning
    Shi, Longxiang
    Li, Shijian
    Zheng, Qian
    Yao, Min
    Pan, Gang
    [J]. IEEE ACCESS, 2020, 8 : 128809 - 128818
  • [2] Deep Reinforcement Learning Autoencoder with Noisy Feedback
    Goutay, Mathieu
    Aoudia, Faycal Ait
    Hoydis, Jakob
    [J]. 17TH INTERNATIONAL SYMPOSIUM ON MODELING AND OPTIMIZATION IN MOBILE, AD HOC, AND WIRELESS NETWORKS (WIOPT 2019), 2019, : 346 - 351
  • [3] Enhancing wound healing through deep reinforcement learning for optimal therapeutics
    Lu, Fan
    Zlobina, Ksenia
    Rondoni, Nicholas A.
    Teymoori, Sam
    Gomez, Marcella
    [J]. ROYAL SOCIETY OPEN SCIENCE, 2024, 11 (07):
  • [4] Enhancing HVAC control systems through transfer learning with deep reinforcement learning agents
    Kadamala, Kevlyn
    Chambers, Des
    Barrett, Enda
    [J]. SMART ENERGY, 2024, 13
  • [5] Deep Reinforcement Learning with Noisy Exploration for Autonomous Driving
    Li, Ruyang
    Zhang, Yaqiang
    Zhao, Yaqian
    Wei, Hui
    Xu, Zhe
    Zhao, Kun
    [J]. PROCEEDINGS OF 2022 THE 6TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND SOFT COMPUTING, ICMLSC 20222, 2022, : 8 - 14
  • [6] Enhancing efficiency and propulsion in bio-mimetic robotic fish through end-to-end deep reinforcement learning
    Cui, Xinyu
    Sun, Boai
    Zhu, Yi
    Yang, Ning
    Zhang, Haifeng
    Cui, Weicheng
    Fan, Dixia
    Wang, Jun
    [J]. PHYSICS OF FLUIDS, 2024, 36 (03)
  • [7] Enhancing Vehicular Cooperative Downloading with Continuous Seeding through Deep Reinforcement Learning
    Niebisch, Michael
    Pfaller, Daniel
    Djanatliev, Anatoli
    [J]. 2023 IEEE LATIN-AMERICAN CONFERENCE ON COMMUNICATIONS, LATINCOM, 2023,
  • [8] Enhancing gas detection-based swarming through deep reinforcement learning
    Lee, Sangmin
    Park, Seongjoon
    Kim, Hwangnam
    [J]. JOURNAL OF SUPERCOMPUTING, 2022, 78 (13): : 14794 - 14812
  • [9] Enhancing gas detection-based swarming through deep reinforcement learning
    Sangmin Lee
    Seongjoon Park
    Hwangnam Kim
    [J]. The Journal of Supercomputing, 2022, 78 : 14794 - 14812
  • [10] Enhancing Deep Reinforcement Learning with Executable Specifications
    Yerushalmi, Raz
    [J]. 2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING: COMPANION PROCEEDINGS, ICSE-COMPANION, 2023, : 213 - 217