Testing probabilistic equivalence through Reinforcement Learning

被引:1
|
作者
Desharnais, Josee [1 ]
Laviolette, Francois [1 ]
Zhioua, Sami [2 ]
机构
[1] Univ Laval, Quebec City, PQ G1K 7P4, Canada
[2] King Fahd Univ Petr & Minerals, ICS, Dhahran 31261, Saudi Arabia
关键词
Verification; Stochastic systems; Markov processes; Distance; Divergence; Reinforcement Learning; Testing; Equivalence relations; BISIMULATION; DIFFERENCE;
D O I
10.1016/j.ic.2013.02.002
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Checking if a given system implementation respects its specification is often done by proving that the two are "equivalent". The equivalence is chosen, in particular, for its computability and of course for its meaning, that is, for its adequacy with what is observable from the two systems (implementation and specification). Trace equivalence is easily testable (decidable from interaction), but often considered too weak; in contrast, bisimulation is accepted as the canonical equivalence for interaction, but it is not testable. Richer than an equivalence is a form of distance: it is zero between equivalent systems, and it provides an estimation of their difference if the systems are not equivalent. Our main contribution is to define such a distance in a context where (1) the two systems to be compared have a stochastic behavior; (2) the model of one of them (e.g., the implementation) is unknown, hence our only knowledge is obtained by interacting with it; (3) consequently the target equivalence (observed when distance is zero) must be testable. To overcome the problem that the model is unknown, we use a Reinforcement Learning approach that provides powerful stochastic algorithms that only need to interact with the model. Our second main contribution is a new family of testable equivalences, called K-moment. The weakest of them, 1-moment equivalence, is trace equivalence; as K grows, K-moment equivalences become finer, all remaining, as well as their limit, weaker than bisimulation. We propose a framework to define (and test) a bigger class of testable equivalences: Test-Observation-Equivalences (TOEs), and we show how they can be made coarser or not, by tuning some parameters. (C) 2013 Elsevier Inc. All rights reserved.
引用
收藏
页码:21 / 57
页数:37
相关论文
共 50 条
  • [41] Probabilistic Multi-knowledge Transfer in Reinforcement Learning
    Fernandez, Daniel
    Fernandez, Fernando
    Garcia, Javier
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 471 - 476
  • [42] Frontostriatal development and probabilistic reinforcement learning during adolescence
    DePasque, Samantha
    Galvan, Adriana
    NEUROBIOLOGY OF LEARNING AND MEMORY, 2017, 143 : 1 - 7
  • [43] A Probabilistic Perspective on Risk-sensitive Reinforcement Learning
    Noorani, Erfaun
    Baras, John S.
    2022 AMERICAN CONTROL CONFERENCE, ACC, 2022, : 2697 - 2702
  • [44] Efficient Reinforcement Learning via Probabilistic Trajectory Optimization
    Pan, Yunpeng
    Boutselis, George, I
    Theodorou, Evangelos A.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (11) : 5459 - 5474
  • [45] Reinforcement learning with algorithms from probabilistic structure estimation
    Epperlein, Jonathan P.
    Overko, Roman
    Zhuk, Sergiy
    King, Christopher
    Bouneffouf, Djallel
    Cullen, Andrew
    Shorten, Robert
    AUTOMATICA, 2022, 144
  • [46] Safe Reinforcement Learning via Probabilistic Logic Shields
    Yang, Wen-Chi
    Marra, Giuseppe
    Rens, Gavin
    De Raedt, Luc
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 5739 - 5749
  • [47] Safe Reinforcement Learning via Probabilistic Logic Shields
    Yang, Wen-Chi
    Marra, Giuseppe
    Rens, Gavin
    De Raedt, Luc
    NEURAL-SYMBOLIC LEARNING AND REASONING 2023, NESY 2023, 2023,
  • [48] Efficient Probabilistic Performance Bounds for Inverse Reinforcement Learning
    Brown, Daniel S.
    Niekum, Scott
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 2754 - 2762
  • [49] Probabilistic Constraint for Safety-Critical Reinforcement Learning
    Chen, Weiqin
    Subramanian, Dharmashankar
    Paternain, Santiago
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (10) : 6789 - 6804
  • [50] A Probabilistic Interpretation of Self-Paced Learning with Applications to Reinforcement Learning
    Klink, Pascal
    Abdulsamad, Hany
    Belousov, Boris
    D'Eramo, Carlo
    Peters, Jan
    Pajarinen, Joni
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22