Can recurrent neural networks learn process model structure?

被引:6
|
作者
Peeperkorn, Jari [1 ]
Broucke, Seppe vanden [1 ,2 ]
De Weerdt, Jochen [1 ]
机构
[1] Katholieke Univ Leuven, Res Ctr Informat Syst Engn LIRIS, Leuven, Belgium
[2] Univ Ghent, Dept Business Informat & Operat Management, Ghent, Belgium
关键词
Process mining; Predictive process analytics; LSTM; Fitness; Precision; Generalization;
D O I
10.1007/s10844-022-00765-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Various methods using machine and deep learning have been proposed to tackle different tasks in predictive process monitoring, forecasting for an ongoing case e.g. the most likely next event or suffix, its remaining time, or an outcome-related variable. Recurrent neural networks (RNNs), and more specifically long short-term memory nets (LSTMs), stand out in terms of popularity. In this work, we investigate the capabilities of such an LSTM to actually learn the underlying process model structure of an event log. We introduce an evaluation framework that combines variant-based resampling and custom metrics for fitness, precision and generalization. We evaluate 4 hypotheses concerning the learning capabilities of LSTMs, the effect of overfitting countermeasures, the level of incompleteness in the training set and the level of parallelism in the underlying process model. We confirm that LSTMs can struggle to learn process model structure, even with simplistic process data and in a very lenient setup. Taking the correct anti-overfitting measures can alleviate the problem. However these measures did not present themselves to be optimal when selecting hyperparameters purely on predicting accuracy. We also found that decreasing the amount of information seen by the LSTM during training, causes a sharp drop in generalization and precision scores. In our experiments, we could not identify a relationship between the extent of parallelism in the model and the generalization capability, but they do indicate that the process' complexity might have impact.
引用
收藏
页码:27 / 51
页数:25
相关论文
共 50 条
  • [41] Recurrent neural networks with multi-branch structure
    Graduate School of Information, Production and Systems, Waseda University, 2-7 Hibikino, Kitakyushu-shi Fukuoka 808-0135, Japan
    不详
    IEEJ Trans. Electron. Inf. Syst., 2007, 9 (1430-1435+19):
  • [42] Reproducing chaos by variable structure recurrent neural networks
    Felix, RA
    Sanchez, EN
    Chen, GR
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2004, 15 (06): : 1450 - 1457
  • [43] Finding event structure in time: What recurrent neural networks can tell us about event structure in mind
    Davis, Forrest
    Altmann, Gerry T. M.
    COGNITION, 2021, 213
  • [44] Recurrent networks learn to tell time
    Renart, Alfonso
    NATURE NEUROSCIENCE, 2013, 16 (07) : 772 - 774
  • [45] Recurrent networks learn to tell time
    Alfonso Renart
    Nature Neuroscience, 2013, 16 : 772 - 774
  • [46] Can You Learn an Algorithm? Generalizing from Easy to Hard Problems with Recurrent Networks
    Schwarzschild, Avi
    Borgnia, Eitan
    Gupta, Arjun
    Huang, Furong
    Vishkin, Uzi
    Goldblum, Micah
    Goldstein, Tom
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [47] Can point-cloud based neural networks learn fingerprint variability?
    Sollinger, Dominik
    Jochl, Robert
    Kirchgasser, Simon
    Uhl, Andreas
    PROCEEDINGS OF THE 21ST 2022 INTERNATIONAL CONFERENCE OF THE BIOMETRICS SPECIAL INTEREST GROUP (BIOSIG 2022), 2022, P-329
  • [48] Learn codes: Inventing low-latency codes via recurrent neural networks
    Jiang Y.
    Kim H.
    Asnani H.
    Kannan S.
    Oh S.
    Viswanath P.
    IEEE Journal on Selected Areas in Information Theory, 2020, 1 (01): : 207 - 216
  • [49] LEARN Codes: Inventing low-latency codes via recurrent neural networks
    Jiang, Yihan
    Kim, Hyeji
    Asnani, Himanshu
    Kannan, Sreeram
    Oh, Sewoong
    Viswanath, Pramod
    ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [50] Biologically plausible gated recurrent neural networks for working memory and learning-to-learn
    van den Berg, Alexandra R.
    Roelfsema, Pieter R.
    Bohte, Sander M.
    PLOS ONE, 2024, 19 (12):