Fuzzy inference system learning by reinforcement methods

被引:229
|
作者
Jouffe, L [1 ]
机构
[1] Inst Natl Sci Appl, Dept Comp Sci, IRISA, SODALEC Elect, F-35043 Rennes, France
关键词
Dynamic Programming (DP); fuzzy logic; learning; Markovian Decision Problem (MDP); reinforcement;
D O I
10.1109/5326.704563
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Fuzzy Actor-Critic Learning (FACL) and Fuzzy Q-Learning (FQL) are reinforcement learning methods based on Dynamic Programming (DP) principles. In this paper, they are used to tune online the conclusion part of Fuzzy Inference Systems (FIS), The only information available for learning is the system feedback, which describes in terms of reward and punishment the task the fuzzy agent has to realize. At each time step, the agent receives a reinforcement signal according to the last action it has performed in the previous state. The problem involves optimizing not only the direct reinforcement, but also the total amount of reinforcements the agent can receive in the future. To illustrate the use of these two learning methods, we first applied them to a problem that involves finding a fuzzy controller to drive a boat from one bank to another, across a river with a strong nonlinear current. Then, we used the well-known Cart-Pole Balancing and Mountain-Car problems to be able to compare our methods to other reinforcement learning methods and focus on important characteristic aspects of FACL and FQL, We found that the genericity of our methods allows us to learn every kind of reinforcement learning problem (continuous states, discrete/continuous actions, various type of reinforcement functions). The experimental studies also show the superiority of these methods with respect to the other related methods we can find in the literature.
引用
收藏
页码:338 / 355
页数:18
相关论文
共 50 条
  • [31] An Interpretable Dynamic Inference System Based on Fuzzy Broad Learning
    Zhao, Huimin
    Wu, Yandong
    Deng, Wu
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [32] Online Reinforcement Learning by Bayesian Inference
    Xia, Zhongpu
    Zhao, Dongbin
    2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [33] Fuzzy Wavelet Network with Reinforcement Learning: Application on Underactuated System
    Razo-Zapata, Ivan S.
    Ramos-Velasco, Luis E.
    Ramos Fernandez, Julio C.
    Espejel-Rivera, Maria A.
    Waissman-Vilanova, Julio
    2012 WORLD AUTOMATION CONGRESS (WAC), 2012,
  • [34] Gene Networks Inference by Reinforcement Learning
    Bonini, Rodrigo Cesar
    Martins-, David Correa, Jr.
    ADVANCES IN BIOINFORMATICS AND COMPUTATIONAL BIOLOGY, BSB 2023, 2023, 13954 : 136 - 147
  • [35] Picture inference system: a new fuzzy inference system on picture fuzzy set
    Le Hoang Son
    Pham Van Viet
    Pham Van Hai
    APPLIED INTELLIGENCE, 2017, 46 (03) : 652 - 669
  • [36] Picture inference system: a new fuzzy inference system on picture fuzzy set
    Le Hoang Son
    Pham Van Viet
    Pham Van Hai
    Applied Intelligence, 2017, 46 : 652 - 669
  • [37] A fuzzy-inference-based reinforcement learning method of overtaking decision making for automated vehicles
    Wu, Qiong
    Cheng, Shuo
    Li, Liang
    Yang, Fan
    Meng, Li Jun
    Fan, Zhi Xian
    Liang, Hua Wei
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART D-JOURNAL OF AUTOMOBILE ENGINEERING, 2022, 236 (01) : 75 - 83
  • [38] Fuzzy Inference Methods Applied to the Learning Competence Measure in Dynamic Classifier Selection
    Kurzynski, Marek
    Krysmann, Maciej
    2014 27TH SIBGRAPI CONFERENCE ON GRAPHICS, PATTERNS AND IMAGES (SIBGRAPI), 2014, : 180 - 187
  • [39] On convergence of fuzzy reinforcement learning
    Berenji, HR
    Vengerov, D
    10TH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-3: MEETING THE GRAND CHALLENGE: MACHINES THAT SERVE PEOPLE, 2001, : 618 - 621
  • [40] Some Methods of Fuzzy Conditional Inference and Fuzzy Reasoning
    Reddy, P. Venkats Subba
    2013 INTERNATIONAL CONFERENCE ON FUZZY THEORY AND ITS APPLICATIONS (IFUZZY 2013), 2013, : 61 - 64