Accelerated multi-objective task learning using modified Q-learning algorithm

被引:0
|
作者
Rajamohan, Varun Prakash [1 ]
Jagatheesaperumal, Senthil Kumar [1 ]
机构
[1] Mepco Schlenk Engn Coll, Dept Elect & Commun Engn, Sivakasi, Tamil Nadu, India
关键词
reinforcement learning; Q-learning; robotic manipulator; task learning; distance metric;
D O I
10.1504/IJAHUC.2024.140665
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robots find extensive applications in industry. In recent years, the influence of robots has also increased rapidly in domestic scenarios. The Q-learning algorithm aims to maximise the reward for reaching the goal. This paper proposes a modified version of the Q-learning algorithm, known as Q-learning with scaled distance metric (Q - SD). This algorithm enhances task learning and makes task completion more meaningful. A robotic manipulator (agent) applies the Q - SD algorithm to the task of table cleaning. Using Q - SD, the agent acquires the sequence of steps necessary to accomplish the task while minimising the manipulator's movement distance. We partition the table into grids of different dimensions. The first has a grid count of 3 x 3, and the second has a grid count of 4 x 4. Using the Q - SD algorithm, the maximum success obtained in these two environments was 86% and 59% respectively. Moreover, compared to the conventional Q-learning algorithm, the drop in average distance moved by the agent in these two environments using the Q - SD algorithm was 8.61% and 6.7% respectively.
引用
收藏
页码:28 / 37
页数:10
相关论文
共 50 条
  • [31] Q-Learning Policies for Multi-Agent Foraging Task
    Yogeswaran, M.
    Ponnambalam, S. C.
    TRENDS IN INTELLIGENT ROBOTICS, 2010, 103 : 194 - 201
  • [32] Multi-objective firefly algorithm with hierarchical learning
    Lv, Li
    Zhou, Xiao-Dong
    Kang, Ping
    Fu, Xue-Feng
    Tian, Xiu-Mei
    Journal of Network Intelligence, 2021, 6 (03): : 411 - 427
  • [33] Multi-objective immune algorithm with Baldwinian learning
    Qi, Yutao
    Liu, Fang
    Liu, Meiyun
    Gong, Maoguo
    Jiao, Licheng
    APPLIED SOFT COMPUTING, 2012, 12 (08) : 2654 - 2674
  • [34] Jamming mitigation in cognitive radio networks using a modified Q-learning algorithm
    Slimeni, Feten
    Scheers, Bart
    Chtourou, Zied
    Le Nir, Vincent
    2015 INTERNATIONAL CONFERENCE ON MILITARY COMMUNICATIONS AND INFORMATION SYSTEMS (ICMCIS), 2015,
  • [35] Q-learning based multi-objective immune algorithm for fuzzy flexible job shop scheduling problem considering dynamic disruptions
    Chen, Xiao-long
    Li, Jun-qing
    Xu, Ying
    SWARM AND EVOLUTIONARY COMPUTATION, 2023, 83
  • [36] MULTI-ROBOT COOPERATIVE TRANSPORTATION OF OBJECTS USING MODIFIED Q-LEARNING
    Siriwardana, Pallege Gamini Dilupa
    de Silva, Clarence
    PROCEEDINGS OF THE ASME INTERNATIONAL MECHANICAL ENGINEERING CONGRESS AND EXPOSITION - 2010, VOL 8, PTS A AND B, 2012, : 745 - 753
  • [37] Multi-Target Tracking Using a Swarm of UAVs by Q-learning Algorithm
    Soleymani, Seyed Ahmad
    Goudarzi, Shidrokh
    Liu, Xingchi
    Mihaylova, Lyudmila
    Wang, Wenwu
    Xiao, Pei
    2023 SENSOR SIGNAL PROCESSING FOR DEFENCE CONFERENCE, SSPD, 2023, : 41 - 45
  • [38] Multi-Robot Task Allocation Using Multimodal Multi-Objective Evolutionary Algorithm Based on Deep Reinforcement Learning
    Miao Z.
    Huang W.
    Zhang Y.
    Fan Q.
    Journal of Shanghai Jiaotong University (Science), 2024, 29 (03) : 377 - 387
  • [39] Decomposition based Multi-Objective Evolutionary Algorithm in XCS for Multi-Objective Reinforcement Learning
    Cheng, Xiu
    Browne, Will N.
    Zhang, Mengjie
    2018 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2018, : 622 - 629
  • [40] QMR:Q-learning based Multi-objective optimization Routing protocol for Flying Ad Hoc Networks
    Liu, Jianmin
    Wang, Qi
    He, ChenTao
    Jaffres-Runser, Katia
    Xu, Yida
    Li, Zhenyu
    Xu, YongJun
    COMPUTER COMMUNICATIONS, 2020, 150 : 304 - 316