Can recurrent neural networks learn process model structure?

被引:6
|
作者
Peeperkorn, Jari [1 ]
Broucke, Seppe vanden [1 ,2 ]
De Weerdt, Jochen [1 ]
机构
[1] Katholieke Univ Leuven, Res Ctr Informat Syst Engn LIRIS, Leuven, Belgium
[2] Univ Ghent, Dept Business Informat & Operat Management, Ghent, Belgium
关键词
Process mining; Predictive process analytics; LSTM; Fitness; Precision; Generalization;
D O I
10.1007/s10844-022-00765-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Various methods using machine and deep learning have been proposed to tackle different tasks in predictive process monitoring, forecasting for an ongoing case e.g. the most likely next event or suffix, its remaining time, or an outcome-related variable. Recurrent neural networks (RNNs), and more specifically long short-term memory nets (LSTMs), stand out in terms of popularity. In this work, we investigate the capabilities of such an LSTM to actually learn the underlying process model structure of an event log. We introduce an evaluation framework that combines variant-based resampling and custom metrics for fitness, precision and generalization. We evaluate 4 hypotheses concerning the learning capabilities of LSTMs, the effect of overfitting countermeasures, the level of incompleteness in the training set and the level of parallelism in the underlying process model. We confirm that LSTMs can struggle to learn process model structure, even with simplistic process data and in a very lenient setup. Taking the correct anti-overfitting measures can alleviate the problem. However these measures did not present themselves to be optimal when selecting hyperparameters purely on predicting accuracy. We also found that decreasing the amount of information seen by the LSTM during training, causes a sharp drop in generalization and precision scores. In our experiments, we could not identify a relationship between the extent of parallelism in the model and the generalization capability, but they do indicate that the process' complexity might have impact.
引用
收藏
页码:27 / 51
页数:25
相关论文
共 50 条
  • [21] Neural networks can learn to utilize correlated auxiliary noise
    Ahmadzadegan, Aida
    Simidzija, Petar
    Li, Ming
    Kempf, Achim
    SCIENTIFIC REPORTS, 2021, 11 (01)
  • [22] DYNAMIC PROCESS MODELING WITH RECURRENT NEURAL NETWORKS
    YOU, Y
    NIKOLAOU, M
    AICHE JOURNAL, 1993, 39 (10) : 1654 - 1667
  • [23] Autonomous Process Model Identification using Recurrent Neural Networks and Hyperparameter Optimization
    Mercangoez, Mehmet
    Cortinovis, Andrea
    Schoenborn, Sandro
    IFAC PAPERSONLINE, 2020, 53 (02): : 11614 - 11619
  • [24] Predicting Sequential Design Decisions Using the Function-Behavior-Structure Design Process Model and Recurrent Neural Networks
    Rahman, Molla Hafizur
    Xie, Charles
    Sha, Zhenghui
    JOURNAL OF MECHANICAL DESIGN, 2021, 143 (08)
  • [25] Structure of activity in multiregion recurrent neural networks
    Clark, David G.
    Beiran, Manuel
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2025, 122 (10)
  • [26] Structure and dynamics of random recurrent neural networks
    Berry, Hugues
    Quoy, Mathias
    ADAPTIVE BEHAVIOR, 2006, 14 (02) : 129 - 137
  • [27] The optimal structure of recurrent neural networks for forecasting
    Pattamavorakun, S
    Phien, HN
    Proceedings of the IASTED International Conference on Applied Simulation and Modelling, 2004, : 77 - 82
  • [28] CAN NEURAL NETWORKS COMPETE WITH PROCESS CALCULATIONS
    BLAESI, J
    JENSEN, B
    INTECH, 1992, 39 (12) : 34 - 37
  • [29] Application of recurrent neural networks to model the defluoridation process of hydroxyapatite synthesized by simple methods
    Gao, Zhipeng
    Liu, Cheng
    Yang, Weihong
    SEPARATION AND PURIFICATION TECHNOLOGY, 2023, 305
  • [30] Information encoding by deep neural networks: what can we learn?
    ten Bosch, L.
    Boves, L.
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 1457 - 1461