Perspectives of approximate dynamic programming

被引:53
|
作者
Powell, Warren B. [1 ]
机构
[1] Princeton Univ, Dept Operat Res & Financial Engn, Princeton, NJ 08544 USA
关键词
LEARNING CONTROL; SELECTION; ALGORITHMS; NETWORKS;
D O I
10.1007/s10479-012-1077-6
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
Approximate dynamic programming has evolved, initially independently, within operations research, computer science and the engineering controls community, all searching for practical tools for solving sequential stochastic optimization problems. More so than other communities, operations research continued to develop the theory behind the basic model introduced by Bellman with discrete states and actions, even while authors as early as Bellman himself recognized its limits due to the "curse of dimensionality" inherent in discrete state spaces. In response to these limitations, subcommunities in computer science, control theory and operations research have developed a variety of methods for solving different classes of stochastic, dynamic optimization problems, creating the appearance of a jungle of competing approaches. In this article, we show that there is actually a common theme to these strategies, and underpinning the entire field remains the fundamental algorithmic strategies of value and policy iteration that were first introduced in the 1950's and 60's.
引用
收藏
页码:319 / 356
页数:38
相关论文
共 50 条
  • [31] Intelligent Questionnaires Using Approximate Dynamic Programming
    Logé, Frédéric
    Le Pennec, Erwan
    Amadou-Boubacar, Habiboulaye
    [J]. i-com, 2021, 19 (03) : 227 - 237
  • [32] AMBULANCE REDEPLOYMENT: AN APPROXIMATE DYNAMIC PROGRAMMING APPROACH
    Maxwell, Matthew S.
    Henderson, Shane G.
    Topaloglu, Huseyin
    [J]. PROCEEDINGS OF THE 2009 WINTER SIMULATION CONFERENCE (WSC 2009 ), VOL 1-4, 2009, : 1801 - 1811
  • [33] Empirical Value Iteration for Approximate Dynamic Programming
    Haskell, William B.
    Jain, Rahul
    Kalathil, Dileep
    [J]. 2014 AMERICAN CONTROL CONFERENCE (ACC), 2014, : 495 - 500
  • [34] Approximate Dynamic Programming via Penalty Functions
    Beuchat, Paul N.
    Lygeros, John
    [J]. IFAC PAPERSONLINE, 2017, 50 (01): : 11814 - 11821
  • [35] Sampled fictitious play for approximate dynamic programming
    Epelman, Marina
    Ghate, Archis
    Smith, Robert L.
    [J]. COMPUTERS & OPERATIONS RESEARCH, 2011, 38 (12) : 1705 - 1718
  • [36] An approximate dynamic programming approach for collaborative caching
    Yang, Xinan
    Thomos, Nikolaos
    [J]. ENGINEERING OPTIMIZATION, 2021, 53 (06) : 1005 - 1023
  • [37] ON CONVERGENCE OF APPROXIMATE SOLUTIONS OF A DYNAMIC PROGRAMMING EQUATION
    JANKOWSK.T
    [J]. COLLOQUIUM MATHEMATICUM, 1970, 21 (01) : 149 - &
  • [38] Approximate dynamic programming for ship course control
    Bai, Xuerui
    Yi, Jianqiang
    Zhao, Dongbin
    [J]. ADVANCES IN NEURAL NETWORKS - ISNN 2007, PT 1, PROCEEDINGS, 2007, 4491 : 349 - +
  • [39] Markdown Optimization via Approximate Dynamic Programming
    Özlem Coşgun
    Ufuk Kula
    Cengiz Kahraman
    [J]. International Journal of Computational Intelligence Systems, 2013, 6 : 64 - 78
  • [40] Separable dynamic programming and approximate decomposition methods
    Bertsekas, Dimitri P.
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2007, 52 (05) : 911 - 916