Heuristically accelerated reinforcement learning modularization for multi-agent multi-objective problems

被引:0
|
作者
Leonardo Anjoletto Ferreira
Carlos Henrique Costa Ribeiro
Reinaldo Augusto da Costa Bianchi
机构
[1] Universidade Metodista de São Paulo,
[2] Instituto Tecnológico de Aeronáutica,undefined
[3] Centro Universitário da FEI,undefined
来源
Applied Intelligence | 2014年 / 41卷
关键词
Reinforcement learning; Heuristically accelerated reinforcement learning; Multi-agent systems; Multi-objective problems;
D O I
暂无
中图分类号
学科分类号
摘要
This article presents two new algorithms for finding the optimal solution of a Multi-agent Multi-objective Reinforcement Learning problem. Both algorithms make use of the concepts of modularization and acceleration by a heuristic function applied in standard Reinforcement Learning algorithms to simplify and speed up the learning process of an agent that learns in a multi-agent multi-objective environment. In order to verify performance of the proposed algorithms, we considered a predator-prey environment in which the learning agent plays the role of prey that must escape the pursuing predator while reaching for food in a fixed location. The results show that combining modularization and acceleration using a heuristics function indeed produced simplification and speeding up of the learning process in a complex problem when comparing with algorithms that do not make use of acceleration or modularization techniques, such as Q-Learning and Minimax-Q.
引用
收藏
页码:551 / 562
页数:11
相关论文
共 50 条
  • [1] Heuristically accelerated reinforcement learning modularization for multi-agent multi-objective problems
    Ferreira, Leonardo Anjoletto
    Costa Ribeiro, Carlos Henrique
    da Costa Bianchi, Reinaldo Augusto
    [J]. APPLIED INTELLIGENCE, 2014, 41 (02) : 551 - 562
  • [2] Multi-objective reinforcement learning for designing ethical multi-agent environments
    Rodriguez-Soto, Manel
    Lopez-Sanchez, Maite
    Rodriguez-Aguilar, Juan A.
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023,
  • [3] Multi-objective reinforcement learning for designing ethical multi-agent environments
    Rodriguez-Soto, Manel
    Lopez-Sanchez, Maite
    Rodriguez-Aguilar, Juan A.
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023,
  • [4] Multi-Agent Deep Reinforcement Learning for Resource Allocation in the Multi-Objective HetNet
    Nie, Hongrui
    Li, Shaosheng
    Liu, Yong
    [J]. IWCMC 2021: 2021 17TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE (IWCMC), 2021, : 116 - 121
  • [5] Multi-Objective Dynamic Dispatch Optimisation using Multi-Agent Reinforcement Learning
    Mannion, Patrick
    Mason, Karl
    Devlin, Sam
    Duggan, Jim
    Howley, Enda
    [J]. AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 1345 - 1346
  • [6] Multi-objective optimization of turbine blade profiles based on multi-agent reinforcement learning
    Li, Lele
    Zhang, Weihao
    Li, Ya
    Jiang, Chiju
    Wang, Yufan
    [J]. ENERGY CONVERSION AND MANAGEMENT, 2023, 297
  • [7] A multi-objective multi-agent deep reinforcement learning approach to residential appliance scheduling
    Lu, Junlin
    Mannion, Patrick
    Mason, Karl
    [J]. IET SMART GRID, 2022, 5 (04) : 260 - 280
  • [8] A new Approach on Multi-Agent Multi-Objective Reinforcement Learning based on agents' preferences
    Asl, Zeinab Daavarani
    Derhami, Vali
    Yazdian-Dehkordi, Mehdi
    [J]. 2017 19TH CSI INTERNATIONAL SYMPOSIUM ON ARTIFICIAL INTELLIGENCE AND SIGNAL PROCESSING (AISP), 2017, : 75 - 79
  • [9] Reward shaping for knowledge-based multi-objective multi-agent reinforcement learning
    Mannion, Patrick
    Devlin, Sam
    Duggan, Jim
    Howley, Enda
    [J]. KNOWLEDGE ENGINEERING REVIEW, 2018, 33
  • [10] A Novel Adaptive Weight Selection Algorithm for Multi-Objective Multi-Agent Reinforcement Learning
    Van Moffaert, Kristof
    Brys, Tim
    Chandra, Arjun
    Esterle, Lukas
    Lewis, Peter R.
    Nowe, Ann
    [J]. PROCEEDINGS OF THE 2014 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2014, : 2306 - 2314