AUGMENTING AUTOTELIC AGENTS WITH LARGE LANGUAGE MODELS

被引:0
|
作者
Colas, Cedric [1 ,2 ]
Teodorescu, Laetitia [2 ]
Oudeyer, Pierre-Yves [2 ]
Yuan, Xingdi [3 ]
Cote, Marc-Alexandre [3 ]
机构
[1] MIT, Cambridge, MA 02139 USA
[2] INRIA, Versailles, France
[3] Microsoft Res, Redmond, WA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Humans learn to master open-ended repertoires of skills by imagining and practicing their own goals. This autotelic learning process, literally the pursuit of self-generated (auto) goals (telos), becomes more and more open-ended as the goals become more diverse, abstract and creative. The resulting exploration of the space of possible skills is supported by an inter-individual exploration: goal representations are culturally evolved and transmitted across individuals, in particular using language. Current artificial agents mostly rely on predefined goal representations corresponding to goal spaces that are either bounded (e.g. list of instructions), or unbounded (e.g. the space of possible visual inputs) but are rarely endowed with the ability to reshape their goal representations, to form new abstractions or to imagine creative goals. In this paper, we introduce a language model augmented autotelic agent (LMA3) that leverages a pretrained language model (LM) to support the representation, generation and learning of diverse, abstract, human-relevant goals. The LM is used as an imperfect model of human cultural transmission; an attempt to capture aspects of humans' common-sense, intuitive physics and overall interests. Specifically, it supports three key components of the autotelic architecture: 1) a relabeler that describes the goals achieved in the agent's trajectories, 2) a goal generator that suggests new high-level goals along with their decomposition into subgoals the agent already masters, and 3) reward functions for each of these goals. Without relying on any hand-coded goal representations, reward functions or curriculum, we show that LMA3 agents learn to master a large diversity of skills in a task-agnostic text-based environment.
引用
收藏
页码:205 / 226
页数:22
相关论文
共 50 条
  • [1] Augmenting large language models with chemistry tools
    Bran, Andres M.
    Cox, Sam
    Schilter, Oliver
    Baldassari, Carlo
    White, Andrew D.
    Schwaller, Philippe
    [J]. NATURE MACHINE INTELLIGENCE, 2024, 6 (05) : 525 - 535
  • [2] Augmenting interpretable models with large language models during training
    Chandan Singh
    Armin Askari
    Rich Caruana
    Jianfeng Gao
    [J]. Nature Communications, 14
  • [3] Augmenting interpretable models with large language models during training
    Singh, Chandan
    Askari, Armin
    Caruana, Rich
    Gao, Jianfeng
    [J]. NATURE COMMUNICATIONS, 2023, 14 (01)
  • [4] Artificial intelligence: Augmenting telehealth with large language models
    Snoswell, Centaine L.
    Snoswell, Aaron J.
    Kelly, Jaimon T.
    Caffery, Liam J.
    Smith, Anthony C.
    [J]. JOURNAL OF TELEMEDICINE AND TELECARE, 2023,
  • [5] Evaluating large language models as agents in the clinic
    Mehandru, Nikita
    Miao, Brenda Y.
    Almaraz, Eduardo Rodriguez
    Sushil, Madhumita
    Butte, Atul J.
    Alaa, Ahmed
    [J]. NPJ DIGITAL MEDICINE, 2024, 7 (01)
  • [6] Evaluating large language models as agents in the clinic
    Nikita Mehandru
    Brenda Y. Miao
    Eduardo Rodriguez Almaraz
    Madhumita Sushil
    Atul J. Butte
    Ahmed Alaa
    [J]. npj Digital Medicine, 7
  • [7] Conversational Agents for Dementia using Large Language Models
    Favela, Jesus
    Cruz-Sandoval, Dagoberto
    Parra, Mario O.
    [J]. 2023 MEXICAN INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE, ENC, 2024,
  • [8] Toward Teachable Autotelic Agents
    Sigaud, Olivier
    Akakzia, Ahmed
    Caselles-Dupre, Hugo
    Colas, Cedric
    Oudeyer, Pierre-Yves
    Chetouani, Mohamed
    [J]. IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2023, 15 (03) : 1070 - 1084
  • [9] GeneGPT: augmenting large language models with domain tools for improved access to biomedical information
    Jin, Qiao
    Yang, Yifan
    Chen, Qingyu
    Lu, Zhiyong
    [J]. BIOINFORMATICS, 2024, 40 (02)
  • [10] MSFuzz: Augmenting Protocol Fuzzing with Message Syntax Comprehension via Large Language Models
    Cheng, Mingjie
    Zhu, Kailong
    Chen, Yuanchao
    Yang, Guozheng
    Lu, Yuliang
    Lu, Canju
    [J]. ELECTRONICS, 2024, 13 (13)