Deep reinforcement learning for data-efficient weakly supervised business process anomaly detection

被引:2
|
作者
Elaziz, Eman Abd [1 ]
Fathalla, Radwa [1 ]
Shaheen, Mohamed [1 ]
机构
[1] Arab Acad Sci Technol & Maritime Transport, Coll Comp & Informat Technol, Alexandria, Egypt
关键词
Business process; Deep reinforcement learning; Weakly supervised learning; Variational autoencoder; Long short-term memory; Self-attention; Transformers; Imbalanced data; Process mining; Process discovery; Conformance checking; Balanced accuracy;
D O I
10.1186/s40537-023-00708-5
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The detection of anomalous behavior in business process data is a crucial task for preventing failures that may jeopardize the performance of any organization. Supervised learning techniques are impracticable because of the difficulties of gathering huge amounts of labeled business process anomaly data. For this reason, unsupervised learning techniques and semi-supervised learning approaches trained on entirely labeled normal data have dominated this domain for a long time. However, these methods do not work well because of the absence of prior knowledge of true anomalies. In this study, we propose a deep weakly supervised reinforcement learning-based approach to identify anomalies in business processes by leveraging limited labeled anomaly data. The proposed approach is intended to use a small collection of labeled anomalous data while exploring a huge set of unlabeled data to find new classes of anomalies that are outside the scope of the labeled anomalous data. We created a unique reward function that combined the supervisory signal supplied by a variational autoencoder trained on unlabeled data with the supervisory signal provided by the environment's reward. To further reduce data deficiency, we introduced a sampling method to allow the effective exploration of the unlabeled data and to address the imbalanced data problem, which is a common problem in the anomaly detection field. This approach depends on the proximity between the data samples in the latent space of the variational autoencoder. Furthermore, to efficiently model the sequential nature of business process data and to handle the long-term dependences, we used a long short-term memory network combined with a self-attention mechanism to develop the agent of our reinforcement learning model. Multiple scenarios were used to test the proposed approach on real-world and synthetic datasets. The findings revealed that the proposed approach outperformed five competing approaches by efficiently using the few available anomalous examples.
引用
收藏
页数:35
相关论文
共 50 条
  • [21] Data-efficient deep reinforcement learning with expert demonstration for active flow control
    Zheng, Changdong
    Xie, Fangfang
    Ji, Tingwei
    Zhang, Xinshuai
    Lu, Yufeng
    Zhou, Hongjie
    Zheng, Yao
    [J]. PHYSICS OF FLUIDS, 2022, 34 (11)
  • [22] SDRL: Interpretable and Data-efficient Deep Reinforcement Learning Leveraging Symbolic Planning
    Lyu, Daoming
    Yang, Fangkai
    Liu, Bo
    Gustafson, Steven
    [J]. ELECTRONIC PROCEEDINGS IN THEORETICAL COMPUTER SCIENCE, 2019, (306): : 354 - 354
  • [23] Load Balancing for Communication Networks via Data-Efficient Deep Reinforcement Learning
    Wu, Di
    Kang, Jikun
    Xu, Yi Tian
    Li, Hang
    Li, Jimmy
    Chen, Xi
    Rivkin, Dmitriy
    Jenkin, Michael
    Lee, Taeseop
    Park, Intaik
    Liu, Xue
    Dudek, Gregory
    [J]. 2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,
  • [24] Deep echocardiography: data-efficient supervised and semi-supervised deep learning towards automated diagnosis of cardiac disease
    Ali Madani
    Jia Rui Ong
    Anshul Tibrewal
    Mohammad R. K. Mofrad
    [J]. npj Digital Medicine, 1
  • [25] Deep echocardiography: data-efficient supervised and semi-supervised deep learning towards automated diagnosis of cardiac disease
    Madani, Ali
    Ong, Jia Rui
    Tibrewal, Anshul
    Mofrad, Mohammad R. K.
    [J]. NPJ DIGITAL MEDICINE, 2018, 1
  • [26] EqR: Equivariant Representations for Data-Efficient Reinforcement Learning
    Mondal, Arnab Kumar
    Jain, Vineet
    Siddiqi, Kaleem
    Ravanbakhsh, Siamak
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [27] Data-Efficient Pipeline for Offline Reinforcement Learning with Limited Data
    Nie, Allen
    Flet-Berliac, Yannis
    Jordan, Deon R.
    Steenbergen, William
    Brunskill, Emma
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [28] Data-Efficient Reinforcement Learning for Variable Impedance Control
    Anand, Akhil S.
    Kaushik, Rituraj
    Gravdahl, Jan Tommy
    Abu-Dakka, Fares J.
    [J]. IEEE ACCESS, 2024, 12 : 15631 - 15641
  • [29] BarlowRL: Barlow Twins for Data-Efficient Reinforcement Learning
    Cagatan, Omer Veysel
    Akgun, Baris
    [J]. ASIAN CONFERENCE ON MACHINE LEARNING, VOL 222, 2023, 222
  • [30] Data-Efficient Offline Reinforcement Learning with Approximate Symmetries
    Angelotti, Giorgio
    Drougard, Nicolas
    Chanel, Caroline P. C.
    [J]. AGENTS AND ARTIFICIAL INTELLIGENCE, ICAART 2023, 2024, 14546 : 164 - 186