Novelty and Inductive Generalization in Human Reinforcement Learning

被引:52
|
作者
Gershman, Samuel J. [1 ]
Niv, Yael [2 ,3 ]
机构
[1] MIT, Dept Brain & Cognit Sci, Cambridge, MA 02139 USA
[2] Princeton Univ, Princeton Neurosci Inst, Princeton, NJ 08544 USA
[3] Princeton Univ, Dept Psychol, Princeton, NJ 08544 USA
关键词
Reinforcement learning; Bayesian inference; Exploration-exploitation dilemma; Neophobia; Neophilia; HIERARCHICAL BAYESIAN MODELS; ADAPTIVE NETWORK MODEL; DOPAMINE NEURONS; EXPLORATORY-BEHAVIOR; DECISION-MAKING; REWARD; EXTINCTION; PREDICTION; RAT; CATEGORIZATION;
D O I
10.1111/tops.12138
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
In reinforcement learning (RL), a decision maker searching for the most rewarding option is often faced with the question: What is the value of an option that has never been tried before? One way to frame this question is as an inductive problem: How can I generalize my previous experience with one set of options to a novel option? We show how hierarchical Bayesian inference can be used to solve this problem, and we describe an equivalence between the Bayesian model and temporal difference learning algorithms that have been proposed as models of RL in humans and animals. According to our view, the search for the best option is guided by abstract knowledge about the relationships between different options in an environment, resulting in greater search efficiency compared to traditional RL algorithms previously applied to human cognition. In two behavioral experiments, we test several predictions of our model, providing evidence that humans learn and exploit structured inductive knowledge to make predictions about novel options. In light of this model, we suggest a new interpretation of dopaminergic responses to novelty.
引用
收藏
页码:391 / 415
页数:25
相关论文
共 50 条
  • [31] Inductive biases in theory-based reinforcement learning
    Pouncy, Thomas
    Gershman, Samuel J.
    [J]. COGNITIVE PSYCHOLOGY, 2022, 138
  • [32] Deep Inductive Logic Programming meets Reinforcement Learning
    Bueff, Andreas
    Belle, Vaishak
    [J]. ELECTRONIC PROCEEDINGS IN THEORETICAL COMPUTER SCIENCE, 2023, (385): : 339 - 352
  • [33] A novelty psychological cognition behaviour model based on reinforcement learning
    Liu, Shiyong
    Chang, Ruosong
    Fu, Sang
    [J]. International Journal of Reasoning-based Intelligent Systems, 2019, 11 (01) : 47 - 56
  • [34] The emergence of saliency and novelty responses from Reinforcement Learning principles
    Laurent, Patryk A.
    [J]. NEURAL NETWORKS, 2008, 21 (10) : 1493 - 1499
  • [35] Novelty-Guided Reinforcement Learning via Encoded Behaviors
    Ramamurthy, Rajkumar
    Sifa, Rafet
    Luebbering, Max
    Bauckhage, Christian
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [36] Development of Inductive Generalization
    Fisher, Anna V.
    [J]. CHILD DEVELOPMENT PERSPECTIVES, 2015, 9 (03) : 172 - 177
  • [37] Using Predictive Representations to Improve Generalization in Reinforcement Learning
    Rafols, Eddie J.
    Ring, Mark B.
    Sutton, Richard S.
    Tanner, Brian
    [J]. 19TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI-05), 2005, : 835 - 840
  • [38] Clustering subspace generalization to obtain faster reinforcement learning
    Maryam Hashemzadeh
    Reshad Hosseini
    Majid Nili Ahmadabadi
    [J]. Evolving Systems, 2020, 11 : 89 - 103
  • [39] Experience generalization for multi-agent reinforcement learning
    Pegoraro, R
    Costa, AHR
    Ribeiro, CHC
    [J]. SCCC 2001: XXI INTERNATIONAL CONFERENCE OF THE CHILEAN COMPUTER SCIENCE SOCIETY, PROCEEDINGS, 2001, : 233 - 239
  • [40] Adversarial Discriminative Feature Separation for Generalization in Reinforcement Learning
    Liu, Yong
    Wu, Chunwei
    Xi, Xidong
    Li, Yan
    Cao, Guitao
    Cao, Wenming
    Wang, Hong
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,