Accelerated multi-objective task learning using modified Q-learning algorithm

被引:0
|
作者
Rajamohan, Varun Prakash [1 ]
Jagatheesaperumal, Senthil Kumar [1 ]
机构
[1] Mepco Schlenk Engn Coll, Dept Elect & Commun Engn, Sivakasi, Tamil Nadu, India
关键词
reinforcement learning; Q-learning; robotic manipulator; task learning; distance metric;
D O I
10.1504/IJAHUC.2024.140665
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robots find extensive applications in industry. In recent years, the influence of robots has also increased rapidly in domestic scenarios. The Q-learning algorithm aims to maximise the reward for reaching the goal. This paper proposes a modified version of the Q-learning algorithm, known as Q-learning with scaled distance metric (Q - SD). This algorithm enhances task learning and makes task completion more meaningful. A robotic manipulator (agent) applies the Q - SD algorithm to the task of table cleaning. Using Q - SD, the agent acquires the sequence of steps necessary to accomplish the task while minimising the manipulator's movement distance. We partition the table into grids of different dimensions. The first has a grid count of 3 x 3, and the second has a grid count of 4 x 4. Using the Q - SD algorithm, the maximum success obtained in these two environments was 86% and 59% respectively. Moreover, compared to the conventional Q-learning algorithm, the drop in average distance moved by the agent in these two environments using the Q - SD algorithm was 8.61% and 6.7% respectively.
引用
收藏
页码:28 / 37
页数:10
相关论文
共 50 条
  • [1] An online scalarization multi-objective reinforcement learning algorithm: TOPSIS Q-learning
    Mirzanejad, Mohammad
    Ebrahimi, Morteza
    Vamplew, Peter
    Veisi, Hadi
    KNOWLEDGE ENGINEERING REVIEW, 2022, 37 (04):
  • [2] Multi-objective route recommendation method based on Q-learning algorithm
    Yu, Qingying
    Xiao, Zhenxing
    Yang, Feng
    Gong, Shan
    Shi, Gege
    Chen, Chuanming
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 44 (04) : 7009 - 7025
  • [3] Cognitive networks QoS multi-objective strategy based on Q-learning algorithm
    Wang, B. (wangbowx@163.com), 1600, Advanced Institute of Convergence Information Technology, Myoungbo Bldg 3F,, Bumin-dong 1-ga, Seo-gu, Busan, 602-816, Korea, Republic of (07):
  • [4] Evaluating Q-Learning Policies for Multi-objective Foraging Task in a Multi-agent Environment
    Yogeswaran, M.
    Ponnambalam, S. G.
    INTELLIGENT ROBOTICS AND APPLICATIONS, PT II, 2010, 6425 : 587 - 598
  • [5] A Novel Multi-Objective Deep Q-Network: Addressing Immediate and Delayed Rewards in Multi-Objective Q-Learning
    Zhang, Youming
    IEEE Access, 2024, 12 : 144932 - 144949
  • [6] Multi-objective virtual network embedding algorithm based on Q-learning and curiosity-driven
    He, Mengyang
    Zhuang, Lei
    Tian, Shuaikui
    Wang, Guoqing
    Zhang, Kunli
    EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2018,
  • [7] Decomposed Multi-objective Method Based on Q-Learning for Solving Multi-objective Combinatorial Optimization Problem
    Yang, Anju
    Liu, Yuan
    Zou, Juan
    Yang, Shengxiang
    BIO-INSPIRED COMPUTING: THEORIES AND APPLICATIONS, PT 1, BIC-TA 2023, 2024, 2061 : 59 - 73
  • [8] Multi-Objective Hole-Making Sequence Optimization by Genetic Algorithm Based on Q-Learning
    Zhang, Desong
    Chen, Yanjie
    Zhu, Guangyu
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (06): : 1 - 14
  • [9] Multi-objective virtual network embedding algorithm based on Q-learning and curiosity-driven
    Mengyang He
    Lei Zhuang
    Shuaikui Tian
    Guoqing Wang
    Kunli Zhang
    EURASIP Journal on Wireless Communications and Networking, 2018
  • [10] A decomposition-based multi-objective evolutionary algorithm with Q-learning for adaptive operator selection
    Xue, Fei
    Chen, Yuezheng
    Wang, Peiwen
    Ye, Yunsen
    Dong, Jinda
    Dong, Tingting
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (14): : 21229 - 21283