Model-based reinforcement learning under concurrent schedules of reinforcement in rodents

被引:22
|
作者
Huh, Namjung
Jo, Suhyun
Kim, Hoseok
Sul, Jung Hoon
Jung, Min Whan [1 ]
机构
[1] Ajou Univ, Sch Med, Neurobiol Lab, Inst Med Sci, Suwon 443721, South Korea
关键词
ANTERIOR CINGULATE CORTEX; MIXED-STRATEGY GAME; DECISION-MAKING; PREFRONTAL CORTEX; DOPAMINE NEURONS; MATCHING LAW; HUMANS; CHOICE; REPRESENTATION; STRIATUM;
D O I
10.1101/lm.1295509
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Reinforcement learning theories postulate that actions are chosen to maximize a long-term sum of positive outcomes based on value functions, which are subjective estimates of future rewards. In simple reinforcement learning algorithms, value functions are updated only by trial-and-error, whereas they are updated according to the decision-maker's knowledge or model of the environment in model-based reinforcement learning algorithms. To investigate how animals update value functions, we trained rats under two different free-choice tasks. The reward probability of the unchosen target remained unchanged in one task, whereas it increased over time since the target was last chosen in the other task. The results show that goal choice probability increased as a function of the number of consecutive alternative choices in the latter, but not the former task, indicating that the animals were aware of time-dependent increases in arming probability and used this information in choosing goals. In addition, the choice behavior in the latter task was better accounted for by a model-based reinforcement learning algorithm. Our results show that rats adopt a decision-making process that cannot be accounted for by simple reinforcement learning models even in a relatively simple binary choice task, suggesting that rats can readily improve their decision-making strategy through the knowledge of their environments.
引用
收藏
页码:315 / 323
页数:9
相关论文
共 50 条
  • [31] Online Constrained Model-based Reinforcement Learning
    van Niekerk, Benjamin
    Damianou, Andreas
    Rosman, Benjamin
    CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI2017), 2017,
  • [32] Calibrated Model-Based Deep Reinforcement Learning
    Malik, Ali
    Kuleshov, Volodymyr
    Song, Jiaming
    Nemer, Danny
    Seymour, Harlan
    Ermon, Stefano
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [33] MATHEMATICAL-MODEL OF LEARNING UNDER SCHEDULES OF INTERRESPONSE TIME REINFORCEMENT
    AMBLER, S
    JOURNAL OF MATHEMATICAL PSYCHOLOGY, 1973, 10 (04) : 364 - 386
  • [34] CONCURRENT RANDOM INTERVAL SCHEDULES OF REINFORCEMENT
    KILLEEN, P
    SHUMWAY, G
    PSYCHONOMIC SCIENCE, 1971, 25 (03): : 155 - 156
  • [35] Reporting contingencies of reinforcement in concurrent schedules
    Jones, BM
    Davison, M
    JOURNAL OF THE EXPERIMENTAL ANALYSIS OF BEHAVIOR, 1998, 69 (02) : 161 - 183
  • [36] CONCURRENT SCHEDULES OF REINFORCEMENT IN GUINEA PIG
    BURNSTEIN, DD
    WOLFF, PC
    DALEY, MF
    PSYCHOLOGICAL RECORD, 1967, 17 (03): : 379 - +
  • [37] REINFORCEMENT OF EYE MOVEMENT WITH CONCURRENT SCHEDULES
    SCHROEDER, SR
    HOLLAND, JG
    JOURNAL OF THE EXPERIMENTAL ANALYSIS OF BEHAVIOR, 1969, 12 (06) : 897 - +
  • [38] SIGNALED REINFORCEMENT IN MULTIPLE AND CONCURRENT SCHEDULES
    WILKIE, DM
    JOURNAL OF THE EXPERIMENTAL ANALYSIS OF BEHAVIOR, 1973, 20 (01) : 29 - 36
  • [39] Concurrent Schedules of Reinforcement as "Challenges" to Maintenance
    Peterson, Stephanie M.
    Frieder, Jessica E.
    Quigley, Shawn P.
    Kestner, Kathryn M.
    Goyal, Manish
    Smith, Shilo L.
    Dayton, Elizabeth
    Brower-Breitwieser, Carrie
    EDUCATION AND TREATMENT OF CHILDREN, 2017, 40 (01) : 57 - 76
  • [40] Incremental Learning of Planning Actions in Model-Based Reinforcement Learning
    Ng, Jun Hao Alvin
    Petrick, Ronald P. A.
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3195 - 3201