(Reinforcement?) Learning to forage optimally

被引:31
|
作者
Kolling, Nils [1 ]
Akam, Thomas [1 ,2 ]
机构
[1] Univ Oxford, Dept Expt Psychol, Oxford, England
[2] Champalimaud Ctr Unknown, Champalimaud Neurosci Program, Lisbon, Portugal
基金
英国惠康基金;
关键词
MEDIAL PREFRONTAL CORTEX; MODEL-FREE; NEURAL MECHANISMS; DECISION-MAKING; REWARD; HABITS; TIME; STRIATUM; PREDICTION; BEHAVIOR;
D O I
10.1016/j.conb.2017.08.008
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Foraging effectively is critical to the survival of all animals and this imperative is thought to have profoundly shaped brain evolution. Decisions made by foraging animals often approximate optimal strategies, but the learning and decision mechanisms generating these choices remain poorly understood. Recent work with laboratory foraging tasks in humans suggest their behaviour is poorly explained by model free reinforcement learning, with simple heuristic strategies better describing behaviour in some tasks, and in others evidence of prospective prediction of the future state of the environment. We suggest that model-based average reward reinforcement learning may provide a common framework for understanding these apparently divergent foraging strategies.
引用
收藏
页码:162 / 169
页数:8
相关论文
共 50 条
  • [31] Optimally learning social networks with activations and suppressions
    Angluin, Dana
    Aspnes, James
    Reyzin, Lev
    [J]. THEORETICAL COMPUTER SCIENCE, 2010, 411 (29-30) : 2729 - 2740
  • [32] Learning to optimally schedule internet banner advertisements
    Abe, N
    Nakamura, A
    [J]. MACHINE LEARNING, PROCEEDINGS, 1999, : 12 - 21
  • [33] Optimally Mitigating Backdoor Attacks in Federated Learning
    Walter, Kane
    Mohammady, Meisam
    Nepal, Surya
    Kanhere, Salil S.
    [J]. IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2024, 21 (04) : 2949 - 2963
  • [34] Optimally Deceiving a Learning Leader in Stackelberg Games
    Birmpas, Georgios
    Gan, Jiarui
    Hollender, Alexandros
    Marmolejo-Cossio, Francisco J.
    Rajgopal, Ninad
    Voudouris, Alexandros A.
    [J]. JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2021, 72 : 507 - 531
  • [35] A Robust and Optimally Pruned Extreme Learning Machine
    Freire, Ananda L.
    Rocha Neto, Ajalmar R.
    [J]. INTELLIGENT SYSTEMS DESIGN AND APPLICATIONS (ISDA 2016), 2017, 557 : 88 - 98
  • [36] The Duality of Utility and Information in Optimally Learning Systems
    Belavkin, Roman V.
    [J]. PROCEEDINGS OF THE 2008 7TH IEEE INTERNATIONAL CONFERENCE ON CYBERNETIC INTELLIGENT SYSTEMS, 2008, : 37 - 42
  • [37] Optimally deceiving a learning leader in stackelberg games
    Birmpas, Georgios
    Gan, Jiarui
    Hollender, Alexandros
    Marmolejo-Cossío, Francisco J.
    Rajgopal, Ninad
    Voudouris, Alexandros A.
    [J]. 1600, AI Access Foundation (72): : 507 - 531
  • [38] Bridging Reinforcement Learning and Creativity: Implementing Reinforcement Learning in Processing
    Luo, Jieliang
    Green, Sam
    [J]. SA'18: SIGGRAPH ASIA 2018 COURSES, 2018,
  • [39] Two Steps Reinforcement Learning in Continuous Reinforcement Learning Tasks
    Lopez-Bueno, Ivan
    Garcia, Javier
    Fernandez, Fernando
    [J]. BIO-INSPIRED SYSTEMS: COMPUTATIONAL AND AMBIENT INTELLIGENCE, PT 1, 2009, 5517 : 577 - 584
  • [40] Curriculum Learning in Reinforcement Learning
    Narvekar, Sanmit
    [J]. AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 1528 - 1529