Optimized control for human-multi-robot collaborative manipulation via multi-player Q-learning

被引:8
|
作者
Liu, Xing [1 ,2 ]
Huang, Panfeng [1 ,2 ]
Ge, Shuzhi Sam [3 ,4 ]
机构
[1] Northwestern Polytech Univ, Sch Astronaut, Res Ctr Intelligent Robot, Xian 710072, Peoples R China
[2] Northwestern Polytech Univ, Sch Astronaut, Natl Key Lab Aerosp Flight Dynam, Xian 710072, Peoples R China
[3] Natl Univ Singapore, Dept Elect & Comp Engn, Singapore 117576, Singapore
[4] Qingdao Univ, Inst Future, Qingdao 266071, Peoples R China
基金
中国国家自然科学基金;
关键词
ADAPTATION; GAME;
D O I
10.1016/j.jfranklin.2021.03.017
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, optimized interaction control is investigated for human-multi-robot collaboration con-trol problems, which cannot be described by the traditional impedance controller. To realize global optimized interaction performance, the multi-player non-zero sum game theory is employed to obtain the optimized interaction control of each robot agent. Regarding the game strategies, Nash equilibrium strategy is utilized in this paper. In human-multi-robot collaboration problems, the dynamics parameters of the human arm and the manipulated object are usually unknown. To obviate the dependence on these parameters, the multi-player Q-learning method is employed. Moreover, for the human-multi-robot collaboration problem, the optimized solution is difficult to resolve due to the existence of the desired reference position. A multi-player Nash Q-learning algorithm considering the desired reference position is proposed to deal with the problem. The validity of the proposed method is verified through simulation studies. (C) 2021 The Franklin Institute. Published by Elsevier Ltd. All rights reserved.
引用
收藏
页码:5639 / 5658
页数:20
相关论文
共 50 条
  • [1] H∞ Control for Discrete-Time Multi-Player Systems via Off-Policy Q-Learning
    Li, Jinna
    Xiao, Zhenfei
    IEEE ACCESS, 2020, 8 (08): : 28831 - 28846
  • [2] Optimized Control for Human-Multi-Robot Collaboration via Multi-Agent Adaptive Dynamic Programming
    Liu, Xing
    Ge, Shuzhi Sam
    IFAC PAPERSONLINE, 2020, 53 (02): : 9207 - 9212
  • [3] Off-Policy Q-Learning for Anti-Interference Control of Multi-Player Systems
    Li, Jinna
    Xiao, Zhenfei
    Chai, Tianyou
    Lewis, Frank L.
    Jagannathan, Sarangapani
    IFAC PAPERSONLINE, 2020, 53 (02): : 9189 - 9194
  • [4] Fuzzy Q-learning for a multi-player non-cooperative repeated game
    Ishibuchi, H
    Nakashima, T
    Miyamoto, H
    Oh, CH
    PROCEEDINGS OF THE SIXTH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS I - III, 1997, : 1573 - 1579
  • [5] Min-Max Q-learning for multi-player pursuit-evasion games
    Selvakumar, Jhanani
    Bakolas, Efstathios
    NEUROCOMPUTING, 2022, 475 : 1 - 14
  • [6] Discrete-Time Multi-Player Games Based on Off-Policy Q-Learning
    Li, Jinna
    Xiao, Zhenfei
    Li, Ping
    IEEE ACCESS, 2019, 7 : 134647 - 134659
  • [7] Learning Queuing Strategies in Human-Multi-Robot Interaction
    Raeissi, Masoume M.
    Farinelli, Alessandro
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 2207 - 2209
  • [8] Reinforcement Q-learning and Optimal Tracking Control of Unknown Discrete-time Multi-player Systems Based on Game Theory
    Zhao, Jin-Gang
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2024, 22 (05) : 1751 - 1759
  • [9] Output Feedback H∞ Control for Linear Discrete-Time Multi-Player Systems With Multi-Source Disturbances Using Off-Policy Q-Learning
    Xiao, Zhenfei
    Li, Jinna
    Li, Ping
    IEEE ACCESS, 2020, 8 : 208938 - 208951
  • [10] A Control Barrier Function Approach to Human-multi-robot Safe Interaction
    Lippi, Martina
    Marino, Alessandro
    2021 29TH MEDITERRANEAN CONFERENCE ON CONTROL AND AUTOMATION (MED), 2021, : 604 - 609