Evolution of reinforcement learning in uncertain environments: Emergence of risk-aversion and matching

被引:0
|
作者
Niv, Yael [1 ]
Joel, Daphna [1 ]
Meilijson, Isaac
Ruppin, Eytan
机构
[1] Tel Aviv Univ, Dept Psychol, IL-69978 Tel Aviv, Israel
来源
ADVANCES IN ARTIFICIAL LIFE | 2001年 / 2159卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning (RL) is a fundamental process by which organisms learn to achieve a goal from interactions with the environment. Using Artificial Life techniques we derive (near-)optimal neuronal learning rules in a simple neural network model of decision-making in simulated bumblebees foraging for nectar. The resulting networks exhibit efficient RL, allowing the bees to respond rapidly to changes in reward contingencies. The evolved synaptic plasticity dynamics give rise to varying exploration/exploitation levels from which emerge the well-documented foraging strategies of risk aversion and probability matching. These are shown to be a direct result of optimal RL, providing a biologically founded, parsimonious and novel explanation for these behaviors. Our results are corroborated by a rigorous mathematical analysis and by experiments in mobile robots.
引用
收藏
页码:252 / 261
页数:10
相关论文
共 46 条