iEnsemble: A Framework for Committee Machine Based on Multiagent Systems with Reinforcement Learning

被引:1
|
作者
Uber Junior, Arnoldo [1 ]
de Freitas Filho, Paulo Jose [1 ]
Silveira, Ricardo Azambuja [1 ]
Costa e Lima, Mariana Dehon [1 ]
Reitz, Rodolfo Wilvert [1 ]
机构
[1] Fed Univ Santa Catarina UFSC, Postgrad Program Comp Sci PPGCC, Florianopolis, SC, Brazil
关键词
Committee machine; Ensemble; Multiagent Systems; Reinforcement learning; ENSEMBLES;
D O I
10.1007/978-3-319-62428-0_6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Machine Learning is one of the areas of Artificial Intelligence whose objective is the development of computational techniques for knowledge and building systems able to acquire knowledge automatically. One of the main challenges of learning algorithms is to maximize generalization. Thus the board machine, or a combination of more of a learning machine approach known in literature with the denomination ensemble along with the theory agents, become a promising alternative in this challenge. In this context, this research proposes the iEnsemble framework, which aims to provide a model of the ensemble through a multi-agent system architecture, where generalization, combination and learning are made through agents, through the performance of their respective roles. In the proposal, the agents follow each their life cycle and also perform the iStacking algorithm. This algorithm is based on Stacking method, which uses the reinforcement learning to define the result of the Ensemble. To validate the initial proposal of the framework, some experiments have been performed and the results obtained and limitations are presented.
引用
收藏
页码:65 / 80
页数:16
相关论文
共 50 条
  • [21] Adaptive Multiagent Model Based on Reinforcement Learning for Distributed Generation Systems
    Divenyi, Daniel
    Dan, Andras
    2012 23RD INTERNATIONAL WORKSHOP ON DATABASE AND EXPERT SYSTEMS APPLICATIONS (DEXA), 2012, : 303 - 307
  • [22] Multiagent Reinforcement Social Learning toward Coordination in Cooperative Multiagent Systems
    Hao, Jianye
    Leung, Ho-Fung
    Ming, Zhong
    ACM TRANSACTIONS ON AUTONOMOUS AND ADAPTIVE SYSTEMS, 2015, 9 (04)
  • [23] N-learning: A reinforcement learning paradigm for multiagent systems
    Mansfield, M
    Collins, JJ
    Eaton, M
    Collins, T
    AI 2005: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2005, 3809 : 684 - 694
  • [24] Dispatching Policy Optimizing of Cruise Taxi in a Multiagent‑Based Deep Reinforcement Learning Framework
    Ma X.
    Wuhan Daxue Xuebao (Xinxi Kexue Ban)/Geomatics and Information Science of Wuhan University, 2023, 48 (12): : 2108
  • [25] A Decentralized Communication Framework Based on Dual-Level Recurrence for Multiagent Reinforcement Learning
    Li, Xuesi
    Li, Jingchen
    Shi, Haobin
    Hwang, Kao-Shing
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (02) : 640 - 649
  • [26] Blockchain-Based Distributed Multiagent Reinforcement Learning for Collaborative Multiobject Tracking Framework
    Shen, Jiahao
    Sheng, Hao
    Wang, Shuai
    Cong, Ruixuan
    Yang, Da
    Zhang, Yang
    IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (03) : 778 - 788
  • [27] MULTIAGENT COORDINATION SYSTEMS BASED ON NEURO-FUZZY MODELS WITH REINFORCEMENT LEARNING
    Mendoza, Leonardo Forero
    Batista, Evelyn
    de Mello, Harold Dias
    Pacheco, Marco Aurelio
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, : 931 - 937
  • [28] A Proactive Eavesdropping Game in MIMO Systems Based on Multiagent Deep Reinforcement Learning
    Guo, Delin
    Ding, Hui
    Tang, Lan
    Zhang, Xinggan
    Yang, Lvxi
    Liang, Ying-Chang
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (11) : 8889 - 8904
  • [29] Beyond Reinforcement Learning and Local View in Multiagent Systems
    Bazzan, Ana L. C.
    KUNSTLICHE INTELLIGENZ, 2014, 28 (03): : 179 - 189
  • [30] Reinforcement Learning With Task Decomposition for Cooperative Multiagent Systems
    Sun, Changyin
    Liu, Wenzhang
    Dong, Lu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (05) : 2054 - 2065