A Procedural Constructive Learning Mechanism with Deep Reinforcement Learning for Cognitive Agents

被引:0
|
作者
Rossi, Leonardo de Lellis [1 ,4 ]
Rohmer, Eric [1 ,4 ]
Costa, Paula Dornhofer Paro [1 ,4 ]
Colombini, Esther Luna [2 ,4 ]
Simoes, Alexandre da Silva [3 ,4 ]
Gudwin, Ricardo Ribeiro [1 ,4 ]
机构
[1] Univ Estadual Campinas, Fac Engn Elect & Comp FEEC, Unicamp, Campinas, Brazil
[2] Univ Estadual Campinas, Inst Comp IC, Unicamp, Campinas, Brazil
[3] Univ Estadual Paulista Unesp, Dept Engn Controle & Automacao DECA, Inst Ciencia & Tecnol Sorocaba ICTS, Campus Sorocaba, Sorocaba, SP, Brazil
[4] Univ Estadual Campinas, Hub Artificial Intelligence & Cognit Architectures, Campinas, Brazil
关键词
Cognitive architecture; Neural networks; Deep reinforcement learning; Developmental robotics; CONSCIOUSNESS; DESIDERATA; ROBOT;
D O I
10.1007/s10846-024-02064-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advancements in AI and deep learning have created a growing demand for artificial agents capable of performing tasks within increasingly complex environments. To address the challenges associated with continuous learning constraints and knowledge capacity in this context, cognitive architectures inspired by human cognition have gained significance. This study contributes to existing research by introducing a cognitive-attentional system employing a constructive neural network-based learning approach for continuous acquisition of procedural knowledge. We replace an incremental tabular Reinforcement Learning algorithm with a constructive neural network deep reinforcement learning mechanism for continuous sensorimotor knowledge acquisition, thereby enhancing the overall learning capacity. The primary emphasis of this modification centers on optimizing memory utilization and reducing training time. Our study presents a learning strategy that amalgamates deep reinforcement learning with procedural learning, mirroring the incremental learning process observed in human sensorimotor development. This approach is embedded within the CONAIM cognitive-attentional architecture, leveraging the cognitive tools of CST. The proposed learning mechanism allows the model to dynamically create and modify elements in its procedural memory, facilitating the reuse of previously acquired functions and procedures. Additionally, it equips the model with the capability to combine learned elements to effectively adapt to complex scenarios. A constructive neural network was employed, initiating with an initial hidden layer comprising one neuron. However, it possesses the capacity to adapt its internal architecture in response to its performance in procedural and sensorimotor learning tasks, inserting new hidden layers or neurons. Experimentation conducted through simulations involving a humanoid robot demonstrates the successful resolution of tasks that were previously unsolved through incremental knowledge acquisition. Throughout the training phase, the constructive agent achieved a minimum of 40% greater rewards and executed 8% more actions when compared to other agents. In the subsequent testing phase, the constructive agent exhibited a 15% increase in the number of actions performed in contrast to its counterparts.
引用
收藏
页数:25
相关论文
共 50 条
  • [31] Cognitive Modeling: From GOMS to Deep Reinforcement Learning
    Jokinen, Jussi P. P.
    Oulasvirta, Antti
    Howes, Andrew
    EXTENDED ABSTRACTS OF THE 2024 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2024, 2024,
  • [32] Cognitive Modelling: From GOMS to Deep Reinforcement Learning
    Jokinen, Jussi P. P.
    Oulasvirta, Antti
    Howes, Andrew
    EXTENDED ABSTRACTS OF THE 2022 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2022, 2022,
  • [33] Learning to Teach Reinforcement Learning Agents
    Fachantidis, Anestis
    Taylor, Matthew
    Vlahavas, Ioannis
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2019, 1 (01): : 21 - 42
  • [34] Reinforcement Learning Agents
    C. Ribeiro
    Artificial Intelligence Review, 2002, 17 : 223 - 250
  • [35] Agents and reinforcement learning
    Harlequin's Adaptive Systems Group
    Dr Dobb's J Software Tools Prof Program, 3 (3pp):
  • [36] Reinforcement learning agents
    Ribeiro, C
    ARTIFICIAL INTELLIGENCE REVIEW, 2002, 17 (03) : 223 - 250
  • [37] Agents and reinforcement learning
    Singh, S
    Norvig, P
    Cohn, D
    DR DOBBS JOURNAL, 1997, 22 (03): : 28 - +
  • [38] Procedural Material Generation with Reinforcement Learning
    Li, Beichen
    Hu, Yiwei
    Guerrero, Paul
    Hasan, Milos
    Shi, Liang
    Deschaintre, Valentin
    Matusik, Wojciech
    ACM TRANSACTIONS ON GRAPHICS, 2024, 43 (06):
  • [39] Residual Network for Deep Reinforcement Learning with Attention Mechanism
    Zhu, Hanhua
    Kaneko, Tomoyuki
    JOURNAL OF INFORMATION SCIENCE AND ENGINEERING, 2021, 37 (03) : 517 - 533
  • [40] Cognitive procedural learning in amnesia
    Schmidtke, K
    Handschu, R
    Vollmer, H
    BRAIN AND COGNITION, 1996, 32 (03) : 441 - 467