Counterexample Explanation by Learning Small Strategies in Markov Decision Processes

被引:15
|
作者
Brazdil, Tomas [1 ]
Chatterjee, Krishnendu [2 ]
Chmelik, Martin [2 ]
Fellner, Andreas [2 ]
Kretinsky, Jan [2 ]
机构
[1] Masaryk Univ, Brno, Czech Republic
[2] IST Austria, Klosterneuburg, Austria
来源
关键词
MODEL CHECKING; ABSTRACTION; MTBDDS;
D O I
10.1007/978-3-319-21690-4_10
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
For deterministic systems, a counterexample to a property can simply be an error trace, whereas counterexamples in probabilistic systems are necessarily more complex. For instance, a set of erroneous traces with a sufficient cumulative probability mass can be used. Since these are too large objects to understand and manipulate, compact representations such as subchains have been considered. In the case of probabilistic systems with non-determinism, the situation is even more complex. While a subchain for a given strategy (or scheduler, resolving non-determinism) is a straightforward choice, we take a different approach. Instead, we focus on the strategy itself, and extract the most important decisions it makes, and present its succinct representation. The key tools we employ to achieve this are (1) introducing a concept of importance of a state w.r.t. the strategy, and (2) learning using decision trees. There are three main consequent advantages of our approach. Firstly, it exploits the quantitative information on states, stressing the more important decisions. Secondly, it leads to a greater variability and degree of freedom in representing the strategies. Thirdly, the representation uses a self-explanatory data structure. In summary, our approach produces more succinct and more explainable strategies, as opposed to e.g. binary decision diagrams. Finally, our experimental results show that we can extract several rules describing the strategy even for very large systems that do not fit in memory, and based on the rules explain the erroneous behaviour.
引用
收藏
页码:158 / 177
页数:20
相关论文
共 50 条
  • [1] SOS: Safe, Optimal and Small Strategies for Hybrid Markov Decision Processes
    Ashok, Pranav
    Kretinsky, Jan
    Larsen, Kim Guldstrand
    Le Coeent, Adrien
    Taankvist, Jakob Haahr
    Weininger, Maximilian
    [J]. QUANTITATIVE EVALUATION OF SYSTEMS (QEST 2019), 2019, 11785 : 147 - 164
  • [2] COUNTEREXAMPLE IN CONTINUOUS MARKOV DECISION CHAINS
    LEMBERSKY, MR
    OTT, ML
    [J]. MANAGEMENT SCIENCE SERIES A-THEORY, 1974, 21 (03): : 358 - 359
  • [3] A Counterexample-Guided Abstraction-Refinement Framework for Markov Decision Processes
    Chadha, Rohit
    Viswanathan, Mahesh
    [J]. ACM TRANSACTIONS ON COMPUTATIONAL LOGIC, 2010, 12 (01)
  • [4] Learning to Collaborate in Markov Decision Processes
    Radanovic, Goran
    Devidze, Rati
    Parkes, David C.
    Singla, Adish
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [5] Learning in Constrained Markov Decision Processes
    Singh, Rahul
    Gupta, Abhishek
    Shroff, Ness B.
    [J]. IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2023, 10 (01): : 441 - 453
  • [6] Online Learning in Kernelized Markov Decision Processes
    Chowdhury, Sayak Ray
    Gopalan, Aditya
    [J]. 22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [7] Blackwell Online Learning for Markov Decision Processes
    Li, Tao
    Peng, Guanze
    Zhu, Quanyan
    [J]. 2021 55TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2021,
  • [8] Learning Factored Markov Decision Processes with Unawareness
    Innes, Craig
    Lascarides, Alex
    [J]. 35TH UNCERTAINTY IN ARTIFICIAL INTELLIGENCE CONFERENCE (UAI 2019), 2020, 115 : 123 - 133
  • [9] Bayesian Learning of Noisy Markov Decision Processes
    Singh, Sumeetpal S.
    Chopin, Nicolas
    Whiteley, Nick
    [J]. ACM TRANSACTIONS ON MODELING AND COMPUTER SIMULATION, 2013, 23 (01):
  • [10] Learning Factored Markov Decision Processes with Unawareness
    Innes, Craig
    Lascarides, Alex
    [J]. AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2030 - 2032