Testing probabilistic equivalence through Reinforcement Learning

被引:1
|
作者
Desharnais, Josee [1 ]
Laviolette, Francois [1 ]
Zhioua, Sami [2 ]
机构
[1] Univ Laval, Quebec City, PQ G1K 7P4, Canada
[2] King Fahd Univ Petr & Minerals, ICS, Dhahran 31261, Saudi Arabia
关键词
Verification; Stochastic systems; Markov processes; Distance; Divergence; Reinforcement Learning; Testing; Equivalence relations; BISIMULATION; DIFFERENCE;
D O I
10.1016/j.ic.2013.02.002
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Checking if a given system implementation respects its specification is often done by proving that the two are "equivalent". The equivalence is chosen, in particular, for its computability and of course for its meaning, that is, for its adequacy with what is observable from the two systems (implementation and specification). Trace equivalence is easily testable (decidable from interaction), but often considered too weak; in contrast, bisimulation is accepted as the canonical equivalence for interaction, but it is not testable. Richer than an equivalence is a form of distance: it is zero between equivalent systems, and it provides an estimation of their difference if the systems are not equivalent. Our main contribution is to define such a distance in a context where (1) the two systems to be compared have a stochastic behavior; (2) the model of one of them (e.g., the implementation) is unknown, hence our only knowledge is obtained by interacting with it; (3) consequently the target equivalence (observed when distance is zero) must be testable. To overcome the problem that the model is unknown, we use a Reinforcement Learning approach that provides powerful stochastic algorithms that only need to interact with the model. Our second main contribution is a new family of testable equivalences, called K-moment. The weakest of them, 1-moment equivalence, is trace equivalence; as K grows, K-moment equivalences become finer, all remaining, as well as their limit, weaker than bisimulation. We propose a framework to define (and test) a bigger class of testable equivalences: Test-Observation-Equivalences (TOEs), and we show how they can be made coarser or not, by tuning some parameters. (C) 2013 Elsevier Inc. All rights reserved.
引用
收藏
页码:21 / 57
页数:37
相关论文
共 50 条
  • [31] Reinforcement learning in VANET penetration testing*
    Garrad, Phillip
    Unnikrishnan, Saritha
    RESULTS IN ENGINEERING, 2023, 17
  • [32] Reinforcement Learning for Intelligent Penetration Testing
    Ghanem, Mohamed C.
    Chen, Thomas M.
    PROCEEDINGS OF THE 2018 SECOND WORLD CONFERENCE ON SMART TRENDS IN SYSTEMS, SECURITY AND SUSTAINABILITY (WORLDS4), 2018, : 185 - 192
  • [33] Formal Specification and Testing for Reinforcement Learning
    Varshosaz, Mahsa
    Ghaffari, Mohsen
    Johnsen, Einar Broch
    Wasowski, Andrzej
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2023, 7 (ICFP): : 125 - 158
  • [34] Testing for Fault Diversity in Reinforcement Learning
    Mazouni, Quentin
    Spieker, Helge
    Gotlieb, Arnaud
    Acher, Mathieu
    PROCEEDINGS OF THE 2024 IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATION OF SOFTWARE TEST, AST 2024, 2024, : 136 - 146
  • [35] A Reinforcement Learning method for Perimetry Testing
    Sznitman, Raphael
    Kucur, Serife
    Marquez-Neila, Pablo
    Abegg, Mathias
    Wolf, Sebastian
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2019, 60 (09)
  • [36] Reinforcement Learning for Android GUI Testing
    Adamo, David
    Khan, Md Khorrom
    Koppula, Sreedevi
    Bryce, Renee
    PROCEEDINGS OF THE 9TH ACM SIGSOFT INTERNATIONAL WORKSHOP ON AUTOMATING TEST CASE DESIGN, SELECTION, AND EVALUATION (A-TEST '18), 2018, : 2 - 8
  • [37] Learning to flock through reinforcement
    Durve, Mihir
    Perumal, Fernando
    Celani, Antonio
    PHYSICAL REVIEW E, 2020, 102 (01)
  • [38] Probabilistic Reinforcement Learning in Adults with Autism Spectrum Disorders
    Solomon, Marjorie
    Smith, Anne C.
    Frank, Michael J.
    Ly, Stanford
    Carter, Cameron S.
    AUTISM RESEARCH, 2011, 4 (02) : 109 - 120
  • [39] Modeling changes in probabilistic reinforcement learning during adolescence
    Xia, Liyu
    Master, Sarah L.
    Eckstein, Maria K.
    Baribault, Beth
    Dahl, Ronald E.
    Wilbrecht, Linda
    Collins, Anne Gabrielle Eva
    PLOS COMPUTATIONAL BIOLOGY, 2021, 17 (07)
  • [40] Associative reinforcement learning using linear probabilistic concepts
    Abe, N
    Long, PM
    MACHINE LEARNING, PROCEEDINGS, 1999, : 3 - 11