Cognitive Reinforcement Learning: An Interpretable Decision-Making for Virtual Driver

被引:0
|
作者
Qi, Hao [1 ]
Hou, Enguang [2 ]
Ye, Peijun [3 ]
机构
[1] South China Univ Technol, Sch Future Technol, Guangzhou 511442, Peoples R China
[2] Shandong Jiaotong Univ, Sch Rail Transportat, Jinan 250357, Peoples R China
[3] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence S, Beijing 100190, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; interpretability; virtual driver; parallel cognition; INTELLIGENCE; VEHICLES; BEHAVIOR;
D O I
10.1109/JRFID.2024.3418649
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The interpretability of decision-making in autonomous driving is crucial for the building of virtual driver, promoting the trust worth of artificial intelligence (AI) and the efficiency of human-machine interaction. However, current data-driven methods such as deep reinforcement learning (DRL) directly acquire driving policies from collected data, where the decision-making process is vague for safety validation. To address this issue, this paper proposes cognitive reinforcement learning that can both simulate the human driver's deliberation and provide interpretability of the virtual driver's behaviors. The new method involves cognitive modeling, reinforcement learning and reasoning path extraction. Experiments on the virtual driving environment indicate that our method can semantically interpret the virtual driver's behaviors. The results show that the proposed cognitive reinforcement learning model combines the interpretability of cognitive models with the learning capability of reinforcement learning, providing a new approach for the construction of trustworthy virtual drivers.
引用
收藏
页码:627 / 631
页数:5
相关论文
共 50 条
  • [21] Driver-like decision-making method for vehicle longitudinal autonomous driving based on deep reinforcement learning
    Gao, Zhenhai
    Yan, Xiangtong
    Gao, Fei
    He, Lei
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART D-JOURNAL OF AUTOMOBILE ENGINEERING, 2022, 236 (13) : 3060 - 3070
  • [22] Why cognitive control matters in learning and decision-making
    Wurm, Franz
    Steinhauser, Marco
    NEUROSCIENCE AND BIOBEHAVIORAL REVIEWS, 2022, 136
  • [23] Negotiable Reinforcement Learning for Pareto Optimal Sequential Decision-Making
    Desai, Nishant
    Critch, Andrew
    Russell, Stuart
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [24] Decision-making models on perceptual uncertainty with distributional reinforcement learning
    Xu, Shuyuan
    Liu, Qiao
    Hu, Yuhui
    Xu, Mengtian
    Hao, Jiachen
    GREEN ENERGY AND INTELLIGENT TRANSPORTATION, 2023, 2 (02):
  • [25] MONEYBARL: EXPLOITING PITCHER DECISION-MAKING USING REINFORCEMENT LEARNING
    Sidhu, Gagan
    Caffo, Brian
    ANNALS OF APPLIED STATISTICS, 2014, 8 (02): : 926 - 955
  • [26] Reinforcement Learning with Uncertainty Estimation for Tactical Decision-Making in Intersections
    Hoel, Carl-Johan
    Tram, Tommy
    Sjoberg, Jonas
    2020 IEEE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2020,
  • [27] A Multiple-Attribute Decision-Making Approach to Reinforcement Learning
    Shi, Haobin
    Xu, Meng
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2020, 12 (04) : 695 - 708
  • [28] Unveiling the Decision-Making Process in Reinforcement Learning with Genetic Programming
    Eberhardinger, Manuel
    Rupp, Florian
    Maucher, Johannes
    Maghsudi, Setareh
    ADVANCES IN SWARM INTELLIGENCE, PT I, ICSI 2024, 2024, 14788 : 349 - 365
  • [29] Intrusion Response Decision-making Method Based on Reinforcement Learning
    Yang, Jun-nan
    Zhang, Hong-qi
    Zhang, Chuan-fu
    2018 INTERNATIONAL CONFERENCE ON COMMUNICATION, NETWORK AND ARTIFICIAL INTELLIGENCE (CNAI 2018), 2018, : 154 - 162
  • [30] Research on Decision-Making in Emotional Agent Based on Reinforcement Learning
    Feng Chao
    Chen Lin
    Jiang Kui
    Wei Zhonglin
    Zhai Bing
    2016 2ND IEEE INTERNATIONAL CONFERENCE ON COMPUTER AND COMMUNICATIONS (ICCC), 2016, : 1191 - 1194