Quantum Deep Reinforcement Learning for Robot Navigation Tasks

被引:0
|
作者
Hohenfeld, Hans [1 ]
Heimann, Dirk [1 ]
Wiebe, Felix [2 ,3 ]
Kirchner, Frank [1 ,2 ,3 ]
机构
[1] Univ Bremen, Robot Res Grp, D-28359 Bremen, Germany
[2] Robot Innovat Ctr RIC, D-28359 Bremen, Germany
[3] German Res Ctr Artificial Intelligence DFKI, D-28359 Bremen, Germany
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Task analysis; Quantum mechanics; Quantum circuit; Deep reinforcement learning; Reinforcement learning; Encoding; autonomous agents; robotics; quantum machine learning; quantum computing;
D O I
10.1109/ACCESS.2024.3417808
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We utilize hybrid quantum deep reinforcement learning to learn navigation tasks for a simple, wheeled robot in simulated environments of increasing complexity. For this, we train parameterized quantum circuits (PQCs) with two different encoding strategies in a hybrid quantum-classical setup as well as a classical neural network baseline with the double deep Q network (DDQN) reinforcement learning algorithm. Quantum deep reinforcement learning (QDRL) has previously been studied in several relatively simple benchmark environments, mainly from the OpenAI gym suite. However, scaling behavior and applicability of QDRL to more demanding tasks closer to real-world problems e.g., from the robotics domain, have not been studied previously. Here, we show that quantum circuits in hybrid quantum-classic reinforcement learning setups are capable of learning optimal policies in multiple robotic navigation scenarios with notably fewer trainable parameters compared to a classical baseline. Across a large number of experimental configurations, we find that the employed quantum circuits outperform the classical neural network baselines when equating for the number of trainable parameters. Yet, the classical neural network consistently showed better results concerning training times and stability, with at least one order of magnitude of trainable parameters more than the best-performing quantum circuits. However, validating the robustness of the learning methods in a large and dynamic environment, we find that the classical baseline produces more stable and better performing policies overall. For the two encoding schemes, we observed better results for consecutively encoding the classical state vector on each qubit compared to encoding each component on a separate qubit. Our findings demonstrate that current hybrid quantum machine-learning approaches can be scaled to simple robotic problems while yielding sufficient results, at least in an idealized simulated setting, but there are yet open questions regarding the application to considerably more demanding tasks. We anticipate that our work will contribute to introducing quantum machine learning in general and quantum deep reinforcement learning in particular to more demanding problem domains and emphasize the importance of encoding techniques for classic data in hybrid quantum-classical settings.
引用
收藏
页码:87217 / 87236
页数:20
相关论文
共 50 条
  • [41] Autonomous Navigation by Mobile Robot with Sensor Fusion Based on Deep Reinforcement Learning
    Ou, Yang
    Cai, Yiyi
    Sun, Youming
    Qin, Tuanfa
    [J]. SENSORS, 2024, 24 (12)
  • [42] Robot Navigation in Crowd Based on Dual Social Attention Deep Reinforcement Learning
    Zeng, Hui
    Hu, Rong
    Huang, Xiaohui
    Peng, Zhiying
    [J]. MATHEMATICAL PROBLEMS IN ENGINEERING, 2021, 2021
  • [43] Learning and Reasoning for Robot Dialog and Navigation Tasks
    Lu, Keting
    Zhang, Shiqi
    Stone, Peter
    Chen, Xiaoping
    [J]. SIGDIAL 2020: 21ST ANNUAL MEETING OF THE SPECIAL INTEREST GROUP ON DISCOURSE AND DIALOGUE (SIGDIAL 2020), 2020, : 107 - 117
  • [44] Obtaining Robust Control and Navigation Policies for Multi-robot Navigation via Deep Reinforcement Learning
    Jestel, Christian
    Surmann, Harmtmut
    Stenzel, Jonas
    Urbann, Oliver
    Brehler, Marius
    [J]. 2021 7TH INTERNATIONAL CONFERENCE ON AUTOMATION, ROBOTICS AND APPLICATIONS (ICARA 2021), 2021, : 48 - 54
  • [45] Indoor Navigation with Deep Reinforcement Learning
    Bakale, Vijayalakshmi A.
    Kumar, Yeshwanth V. S.
    Roodagi, Vivekanand C.
    Kulkarni, Yashaswini N.
    Patil, Mahesh S.
    Chickerur, Satyadhyan
    [J]. PROCEEDINGS OF THE 5TH INTERNATIONAL CONFERENCE ON INVENTIVE COMPUTATION TECHNOLOGIES (ICICT-2020), 2020, : 660 - 665
  • [46] Composite Reinforcement Learning for Social Robot Navigation
    Ciou, Pei-Huai
    Hsiao, Yu-Ting
    Wu, Zong-Ze
    Tseng, Shih-Huan
    Fu, Li-Chen
    [J]. 2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 2553 - 2558
  • [47] Reinforcement Learning for Robot Navigation in Nondeterministic Environments
    Liu, Xiaoyun
    Zhou, Qingrui
    Ren, Hailin
    Sun, Changhao
    [J]. PROCEEDINGS OF 2018 5TH IEEE INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND INTELLIGENCE SYSTEMS (CCIS), 2018, : 615 - 619
  • [48] Hierarchies of Planning and Reinforcement Learning for Robot Navigation
    Woehlke, Jan
    Schmitt, Felix
    van Hoof, Herke
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 10682 - 10688
  • [49] Supervised fuzzy reinforcement learning for robot navigation
    Fathinezhad, Fatemeh
    Derhami, Vali
    Rezaeian, Mehdi
    [J]. APPLIED SOFT COMPUTING, 2016, 40 : 33 - 41
  • [50] Reinforcement learning with function approximation for cooperative navigation tasks
    Melo, Francisco S.
    Ribeiro, M. Isabel
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-9, 2008, : 3321 - +