Interactive Q-learning on heterogeneous agents system for autonomous adaptive interface

被引:0
|
作者
Ishiwaka, Y [1 ]
Yokoi, H [1 ]
Kakazu, Y [1 ]
机构
[1] Hakodate Natl Coll Technol, Dept Informat Engn, Hakodate, Hokkaido 0428501, Japan
关键词
Interactive Q-learning (IQL); POSMDP; heterogeneous multiagent system; Khepera;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Purpose of this system is to adapt the bedridden people who cannot move their body easily, so the simple reinforcement signals are applied. The application is to control the behaviors of Khepera robot, which is a small mobile robot. For the simple reinforcement signals the on-off signals are employed when the operators as the training agent feels discomfort for the behaviors of the learning agent Khepera robot. We proposed the new reinforcement learning method called Interactive Q-learning and the heterogeneous multi agent system. Our multi agent system has three kinds of heterogeneous single agent: Learning agent, Training agent and Interface Agent. The system is hierarchic. There are also three hierarchies. It is impossible to iterate the many episodes and steps to converge the learning which is adopted in general reinforcement learning in simulation world. We show the results of experiments using the Khepera robot for 3 examinees, and discuss how to give the rewards according to each operator and the significance of heterogeneous multi agent system. We confirmed the effectiveness through the some experiments which are to control the behavior of Khepera robot in real world. The convergences of our teaming system are quite quick. Furthermore the importance of the interface agent is indicated. The individual differences for the timing to give the penalties are happened even though all operators are young.
引用
收藏
页码:475 / 484
页数:10
相关论文
共 50 条
  • [1] Interactive Q-learning to adapt autonomous interface
    Ishiwaka, Y
    Takehara, N
    Yokoi, H
    Kakazu, Y
    INTELLIGENT AUTONOMOUS VEHICLES 2001, 2002, : 45 - 50
  • [2] Successful cooperation between heterogeneous fuzzy Q-learning agents
    Bitaghsir, AA
    Moghimi, A
    Lesani, M
    Keramati, MM
    Ahmadabadi, MN
    Arabi, BN
    2004 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN & CYBERNETICS, VOLS 1-7, 2004, : 5579 - 5583
  • [3] Interactive Q-Learning for Quantiles
    Linn, Kristin A.
    Laber, Eric B.
    Stefanski, Leonard A.
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2017, 112 (518) : 638 - 649
  • [4] iqLearn: Interactive Q-Learning in R
    Linn, Kristin A.
    Laber, Eric B.
    Stefanski, Leonard A.
    JOURNAL OF STATISTICAL SOFTWARE, 2015, 64 (01):
  • [5] Interactive model building for Q-learning
    Laber, Eric B.
    Linn, Kristin A.
    Stefanski, Leonard A.
    BIOMETRIKA, 2014, 101 (04) : 831 - 847
  • [6] Adaptive Bases for Q-learning
    Di Castro, Dotan
    Mannor, Shie
    49TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2010, : 4587 - 4593
  • [7] Q-Learning for autonomous vehicle navigation
    Gonzalez-Miranda, Oscar
    Miranda, Luis Antonio Lopez
    Ibarra-Zannatha, Juan Manuel
    2023 XXV ROBOTICS MEXICAN CONGRESS, COMROB, 2023, : 138 - 142
  • [8] Recurrent Deep Multiagent Q-Learning for Autonomous Agents in Future Smart Grid
    Yang, Yaodong
    Hao, Jianye
    Wang, Zan
    Sun, Mingyang
    Strbac, Goran
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 2136 - 2138
  • [9] Q-learning with heterogeneous update strategy
    Tan, Tao
    Xie, Hong
    Feng, Liang
    INFORMATION SCIENCES, 2024, 656