Optimized control for human-multi-robot collaborative manipulation via multi-player Q-learning

被引:8
|
作者
Liu, Xing [1 ,2 ]
Huang, Panfeng [1 ,2 ]
Ge, Shuzhi Sam [3 ,4 ]
机构
[1] Northwestern Polytech Univ, Sch Astronaut, Res Ctr Intelligent Robot, Xian 710072, Peoples R China
[2] Northwestern Polytech Univ, Sch Astronaut, Natl Key Lab Aerosp Flight Dynam, Xian 710072, Peoples R China
[3] Natl Univ Singapore, Dept Elect & Comp Engn, Singapore 117576, Singapore
[4] Qingdao Univ, Inst Future, Qingdao 266071, Peoples R China
基金
中国国家自然科学基金;
关键词
ADAPTATION; GAME;
D O I
10.1016/j.jfranklin.2021.03.017
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, optimized interaction control is investigated for human-multi-robot collaboration con-trol problems, which cannot be described by the traditional impedance controller. To realize global optimized interaction performance, the multi-player non-zero sum game theory is employed to obtain the optimized interaction control of each robot agent. Regarding the game strategies, Nash equilibrium strategy is utilized in this paper. In human-multi-robot collaboration problems, the dynamics parameters of the human arm and the manipulated object are usually unknown. To obviate the dependence on these parameters, the multi-player Q-learning method is employed. Moreover, for the human-multi-robot collaboration problem, the optimized solution is difficult to resolve due to the existence of the desired reference position. A multi-player Nash Q-learning algorithm considering the desired reference position is proposed to deal with the problem. The validity of the proposed method is verified through simulation studies. (C) 2021 The Franklin Institute. Published by Elsevier Ltd. All rights reserved.
引用
收藏
页码:5639 / 5658
页数:20
相关论文
共 50 条
  • [41] Multi-User Mm Wave Beam Tracking via Multi-Agent Deep Q-Learning
    MENG Fan
    HUANG Yongming
    LU Zhaohua
    XIAO Huahua
    ZTE Communications, 2023, 21 (02) : 53 - 60
  • [42] Q-learning Based Multi-robot Box-Pushing with Minimal Switching of Actions
    Wang, Ying
    Lang, Haoxiang
    de Silva, Clarence W.
    2008 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS, VOLS 1-6, 2008, : 640 - +
  • [43] Multi-Player Pursuer Coordination for Nonlinear Reach-Avoid Games in Arbitrary Dimensions via Coverage Control
    Rivera-Ortiz, Phillip
    Diaz-Mercado, Yancy
    Kobilarov, Marin
    2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 2747 - 2753
  • [44] Cooperative Deep Q-Learning With Q-Value Transfer for Multi-Intersection Signal Control
    Ge, Hongwei
    Song, Yumei
    Wu, Chunguo
    Ren, Jiankang
    Tan, Guozhen
    IEEE ACCESS, 2019, 7 : 40797 - 40809
  • [45] Q-learning Algorithm Based Multi-Agent Coordinated Control Method for Microgrids
    Xi, Yuanyuan
    Chang, Liuchen
    Mao, Meiqin
    Jin, Peng
    Hatziargyriou, Nikos
    Xu, Haibo
    2015 9TH INTERNATIONAL CONFERENCE ON POWER ELECTRONICS AND ECCE ASIA (ICPE-ECCE ASIA), 2015, : 1497 - 1504
  • [46] Hierarchically Correlated Equilibrium Q-learning for Multi-area Decentralized Collaborative Reactive Power Optimization
    Tan, Min
    Han, Chuanjia
    Zhang, Xiaoshun
    Guo, Lexin
    Yu, Tao
    CSEE JOURNAL OF POWER AND ENERGY SYSTEMS, 2016, 2 (03): : 65 - 72
  • [47] Collaborative Q-learning Path Planning for Autonomous Robots based on Holonic Multi-Agent System
    Lamini, Chaymaa
    Fathi, Youssef
    Benhlima, Said
    2015 10TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS: THEORIES AND APPLICATIONS (SITA), 2015,
  • [48] Multi-agent Q-Learning control of spacecraft formation flying reconfiguration trajectories
    Kankashvar, Mohammadrasoul
    Bolandi, Hossein
    Mozayani, Nasser
    ADVANCES IN SPACE RESEARCH, 2023, 71 (03) : 1627 - 1643
  • [49] Multi-UAV Formation Maneuvering Control Based on Q-Learning Fuzzy Controller
    Rui, Pang
    2ND IEEE INTERNATIONAL CONFERENCE ON ADVANCED COMPUTER CONTROL (ICACC 2010), VOL. 4, 2010, : 252 - 257
  • [50] Q-Learning Based Failure Detection and Self-Recovery Algorithm for Multi-Robot Domains
    Kayir, Hatice Hilal Ezercan
    ELEKTRONIKA IR ELEKTROTECHNIKA, 2019, 25 (02) : 9 - 13