Novelty and Inductive Generalization in Human Reinforcement Learning

被引:52
|
作者
Gershman, Samuel J. [1 ]
Niv, Yael [2 ,3 ]
机构
[1] MIT, Dept Brain & Cognit Sci, Cambridge, MA 02139 USA
[2] Princeton Univ, Princeton Neurosci Inst, Princeton, NJ 08544 USA
[3] Princeton Univ, Dept Psychol, Princeton, NJ 08544 USA
关键词
Reinforcement learning; Bayesian inference; Exploration-exploitation dilemma; Neophobia; Neophilia; HIERARCHICAL BAYESIAN MODELS; ADAPTIVE NETWORK MODEL; DOPAMINE NEURONS; EXPLORATORY-BEHAVIOR; DECISION-MAKING; REWARD; EXTINCTION; PREDICTION; RAT; CATEGORIZATION;
D O I
10.1111/tops.12138
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
In reinforcement learning (RL), a decision maker searching for the most rewarding option is often faced with the question: What is the value of an option that has never been tried before? One way to frame this question is as an inductive problem: How can I generalize my previous experience with one set of options to a novel option? We show how hierarchical Bayesian inference can be used to solve this problem, and we describe an equivalence between the Bayesian model and temporal difference learning algorithms that have been proposed as models of RL in humans and animals. According to our view, the search for the best option is guided by abstract knowledge about the relationships between different options in an environment, resulting in greater search efficiency compared to traditional RL algorithms previously applied to human cognition. In two behavioral experiments, we test several predictions of our model, providing evidence that humans learn and exploit structured inductive knowledge to make predictions about novel options. In light of this model, we suggest a new interpretation of dopaminergic responses to novelty.
引用
收藏
页码:391 / 415
页数:25
相关论文
共 50 条
  • [1] The effect of novelty on reinforcement learning
    Houillon, A.
    Lorenz, R. C.
    Boehmer, W.
    Rapp, M. A.
    Heinz, A.
    Gallinat, J.
    Obermayer, K.
    [J]. DECISION MAKING: NEURAL AND BEHAVIOURAL APPROACHES, 2013, 202 : 415 - 439
  • [2] INDUCTIVE AND DEDUCTIVE GENERALIZATION IN LEARNING BY ACTION
    GEORGE, C
    [J]. ANNEE PSYCHOLOGIQUE, 1986, 86 (02): : 183 - 200
  • [3] On the Generalization of Representations in Reinforcement Learning
    Le Lan, Charline
    Tu, Stephen
    Oberman, Adam
    Agarwal, Rishabh
    Bellemare, Marc
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [4] Quantifying Generalization in Reinforcement Learning
    Cobbe, Karl
    Klimov, Oleg
    Hesse, Chris
    Kim, Taehoon
    Schulman, John
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [5] Learning skills in reinforcement learning using relative novelty
    Simsek, Ö
    Barto, AG
    [J]. ABSTRACTION, REFORMULATION AND APPROXIMATION, PROCEEDINGS, 2005, 3607 : 367 - 374
  • [6] Learning Dynamics and Generalization in Deep Reinforcement Learning
    Lyle, Clare
    Rowland, Mark
    Dabney, Will
    Kwiatkowksa, Marta
    Gal, Yarin
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [7] Novelty Seeking Multiagent Evolutionary Reinforcement Learning
    Aydeniz, Ayhan Alp
    Loftin, Robert
    Tumer, Kagan
    [J]. PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, GECCO 2023, 2023, : 402 - 410
  • [8] Novelty Detector for Reinforcement Learning Based on Forecasting
    Gregor, Michal
    Spalek, Juraj
    [J]. 2014 IEEE 12TH INTERNATIONAL SYMPOSIUM ON APPLIED MACHINE INTELLIGENCE AND INFORMATICS (SAMI), 2014, : 73 - 78
  • [9] On the Generalization Gap in Reparameterizable Reinforcement Learning
    Wang, Huan
    Zheng, Stephan
    Xiong, Caiming
    Socher, Richard
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [10] On the Importance of Exploration for Generalization in Reinforcement Learning
    Jiang, Yiding
    Kolter, J. Zico
    Raileanu, Roberta
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,