An integrated approach to solving influence diagrams and finite-horizon partially observable decision processes

被引:1
|
作者
Hansen, Eric A. [1 ]
机构
[1] Mississippi State Univ, Dept Comp Sci & Engn, Mississippi State, MS 39762 USA
基金
美国国家科学基金会;
关键词
Influence diagram; Variable elimination; Partially observable Markov decision process; Dynamic programming; Decision-theoretic planning; ELIMINATION; INFERENCE;
D O I
10.1016/j.artint.2020.103431
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We show how to integrate a variable elimination approach to solving influence diagrams with a value iteration approach to solving finite-horizon partially observable Markov decision processes (POMDPs). The integration of these approaches creates a variable elimination algorithm for influence diagrams that has much more relaxed constraints on elimination order, which allows improved scalability in many cases. The new algorithm can also be viewed as a generalization of the value iteration algorithm for POMDPs that solves non-Markovian as well as Markovian problems, in addition to leveraging a factored representation for improved efficiency. The development of a single algorithm that integrates and generalizes both of these classic algorithms, one for influence diagrams and the other for POMDPs, unifies these two approaches to solving Bayesian decision problems in a way that combines their complementary advantages. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页数:47
相关论文
共 50 条