Model-based reinforcement learning under concurrent schedules of reinforcement in rodents

被引:22
|
作者
Huh, Namjung
Jo, Suhyun
Kim, Hoseok
Sul, Jung Hoon
Jung, Min Whan [1 ]
机构
[1] Ajou Univ, Sch Med, Neurobiol Lab, Inst Med Sci, Suwon 443721, South Korea
关键词
ANTERIOR CINGULATE CORTEX; MIXED-STRATEGY GAME; DECISION-MAKING; PREFRONTAL CORTEX; DOPAMINE NEURONS; MATCHING LAW; HUMANS; CHOICE; REPRESENTATION; STRIATUM;
D O I
10.1101/lm.1295509
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Reinforcement learning theories postulate that actions are chosen to maximize a long-term sum of positive outcomes based on value functions, which are subjective estimates of future rewards. In simple reinforcement learning algorithms, value functions are updated only by trial-and-error, whereas they are updated according to the decision-maker's knowledge or model of the environment in model-based reinforcement learning algorithms. To investigate how animals update value functions, we trained rats under two different free-choice tasks. The reward probability of the unchosen target remained unchanged in one task, whereas it increased over time since the target was last chosen in the other task. The results show that goal choice probability increased as a function of the number of consecutive alternative choices in the latter, but not the former task, indicating that the animals were aware of time-dependent increases in arming probability and used this information in choosing goals. In addition, the choice behavior in the latter task was better accounted for by a model-based reinforcement learning algorithm. Our results show that rats adopt a decision-making process that cannot be accounted for by simple reinforcement learning models even in a relatively simple binary choice task, suggesting that rats can readily improve their decision-making strategy through the knowledge of their environments.
引用
收藏
页码:315 / 323
页数:9
相关论文
共 50 条
  • [1] Model-based Reinforcement Learning: A Survey
    Moerland, Thomas M.
    Broekens, Joost
    Plaat, Aske
    Jonker, Catholijn M.
    FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2023, 16 (01): : 1 - 118
  • [2] A survey on model-based reinforcement learning
    Fan-Ming LUO
    Tian XU
    Hang LAI
    Xiong-Hui CHEN
    Weinan ZHANG
    Yang YU
    Science China(Information Sciences), 2024, 67 (02) : 59 - 84
  • [3] Nonparametric model-based reinforcement learning
    Atkeson, CG
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 10, 1998, 10 : 1008 - 1014
  • [4] The ubiquity of model-based reinforcement learning
    Doll, Bradley B.
    Simon, Dylan A.
    Daw, Nathaniel D.
    CURRENT OPINION IN NEUROBIOLOGY, 2012, 22 (06) : 1075 - 1081
  • [5] Multiple model-based reinforcement learning
    Doya, K
    Samejima, K
    Katagiri, K
    Kawato, M
    NEURAL COMPUTATION, 2002, 14 (06) : 1347 - 1369
  • [6] A survey on model-based reinforcement learning
    Luo, Fan-Ming
    Xu, Tian
    Lai, Hang
    Chen, Xiong-Hui
    Zhang, Weinan
    Yu, Yang
    SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (02)
  • [7] Novel reinforcement learning paradigm based on response patterning under interval schedules of reinforcement
    Schifani, Christin
    Sukhanov, Ilya
    Dorofeikova, Mariia
    Bespalov, Anton
    BEHAVIOURAL BRAIN RESEARCH, 2017, 331 : 276 - 281
  • [8] CONCURRENT SCHEDULES OF REINFORCEMENT IN THE CHIMPANZEE
    FERSTER, CB
    SCIENCE, 1957, 125 (3257) : 1090 - 1091
  • [9] Reinforcement Twinning: From digital twins to model-based reinforcement learning
    Schena, Lorenzo
    Marques, Pedro A.
    Poletti, Romain
    Van den Berghe, Jan
    Mendez, Miguel A.
    JOURNAL OF COMPUTATIONAL SCIENCE, 2024, 82
  • [10] Learning to Paint With Model-based Deep Reinforcement Learning
    Huang, Zhewei
    Heng, Wen
    Zhou, Shuchang
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 8708 - 8717