Multi-objective fuzzy Q-learning to solve continuous state-action problems

被引:4
|
作者
Asgharnia, Amirhossein [1 ]
Schwartz, Howard [1 ]
Atia, Mohamed [1 ]
机构
[1] Carleton Univ, Dept Syst & Comp Engn, Ottawa, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Reinforcement learning; Differential games; Q-learning; Multi-objective reinforcement learning;
D O I
10.1016/j.neucom.2022.10.035
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many real world problems are multi-objective. Thus, the need for multi-objective learning and optimiza-tion algorithms is inevitable. Although the multi-objective optimization algorithms are well-studied, the multi-objective learning algorithms have attracted less attention. In this paper, a fuzzy multi-objective reinforcement learning algorithm is proposed, and we refer to it as the multi-objective fuzzy Q-learning (MOFQL) algorithm. The algorithm is implemented to solve a bi-objective reach-avoid game. The majority of the multi-objective reinforcement algorithms proposed address solving problems in the discrete state-action domain. However, the MOFQL algorithm can also handle problems in a contin-uous state-action domain. A fuzzy inference system (FIS) is implemented to estimate the value function for the bi-objective problem. We used a temporal difference (TD) approach to update the fuzzy rules. The proposed method isa multi-policy multi-objective algorithm and can find the non-convex regions of the Pareto front.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:115 / 132
页数:18
相关论文
共 50 条
  • [31] Bidirectional Q-Learning based Multi-objective optimization Routing Protocol for Multi-Destination FANETs
    Xue, Liang
    Tang, Jie
    Zhang, Jiaying
    Hu, Juncheng
    2024 IEEE INTERNATIONAL WORKSHOP ON RADIO FREQUENCY AND ANTENNA TECHNOLOGIES, IWRF&AT 2024, 2024, : 421 - 426
  • [32] Q-learning based multi-objective immune algorithm for fuzzy flexible job shop scheduling problem considering dynamic disruptions
    Chen, Xiao-long
    Li, Jun-qing
    Xu, Ying
    SWARM AND EVOLUTIONARY COMPUTATION, 2023, 83
  • [33] Using Goal Programming Approach to Solve Fuzzy Multi-objective Linear Fractional Programming Problems
    De, P. K.
    Deb, Moumita
    2016 IEEE INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND COMPUTING RESEARCH, 2016, : 922 - 926
  • [34] Fuzzy Programming Approach to Solve Multi-objective Transportation Problem
    Kumar, Sandeep
    Pandey, Diwakar
    PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON SOFT COMPUTING FOR PROBLEM SOLVING (SOCPROS 2011), VOL 1, 2012, 130 : 525 - 533
  • [35] Near-continuous time Reinforcement Learning for continuous state-action spaces
    Croissant, Lorenzo
    Abeille, Marc
    Bouchard, Bruno
    INTERNATIONAL CONFERENCE ON ALGORITHMIC LEARNING THEORY, VOL 237, 2024, 237
  • [36] β-Robustness Approach for Fuzzy Multi-objective Problems
    Bahri, Oumayma
    Ben Amor, Nahla
    Talbi, El-Ghazali
    INFORMATION PROCESSING AND MANAGEMENT OF UNCERTAINTY IN KNOWLEDGE-BASED SYSTEMS, IPMU 2016, PT II, 2016, 611 : 225 - 237
  • [37] Pruning the Way to Reliable Policies: A Multi-Objective Deep Q-Learning Approach to Critical Care
    Shirali A.
    Schubert A.
    Alaa A.
    IEEE Journal of Biomedical and Health Informatics, 2024, 28 (10) : 1 - 13
  • [38] Multi-objective virtual network embedding algorithm based on Q-learning and curiosity-driven
    He, Mengyang
    Zhuang, Lei
    Tian, Shuaikui
    Wang, Guoqing
    Zhang, Kunli
    EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2018,
  • [39] Multimodal transportation routing optimization based on multi-objective Q-learning under time uncertainty
    Zhang, Tie
    Cheng, Jia
    Zou, Yanbiao
    COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (02) : 3133 - 3152
  • [40] Multi-Objective Hole-Making Sequence Optimization by Genetic Algorithm Based on Q-Learning
    Zhang, Desong
    Chen, Yanjie
    Zhu, Guangyu
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (06): : 1 - 14