A Monte-Carlo AIXI Approximation

被引:67
|
作者
Veness, Joel [1 ]
Kee Siong Ng [2 ]
Hutter, Marcus [2 ]
Uther, William [1 ]
Silver, David [3 ]
机构
[1] Univ New S Wales, Sydney, NSW 2052, Australia
[2] Australian Natl Univ, Canberra, ACT 0200, Australia
[3] MIT, Cambridge, MA 02139 USA
基金
澳大利亚研究理事会;
关键词
TREE WEIGHTING METHOD; UNIVERSAL; ALGORITHM;
D O I
10.1613/jair.3125
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper introduces a principled approach for the design of a scalable general reinforcement learning agent. Our approach is based on a direct approximation of AIXI, a Bayesian optimality notion for general reinforcement learning agents. Previously, it has been unclear whether the theory of AIXI could motivate the design of practical algorithms. We answer this hitherto open question in the affirmative, by providing the first computationally feasible approximation to the AIXI agent. To develop our approximation, we introduce a new Monte-Carlo Tree Search algorithm along with an agent-specific extension to the Context Tree Weighting algorithm. Empirically, we present a set of encouraging results on a variety of stochastic and partially observable domains. We conclude by proposing a number of directions for future research.
引用
收藏
页码:95 / 142
页数:48
相关论文
共 50 条