Can recurrent neural networks learn process model structure?

被引:6
|
作者
Peeperkorn, Jari [1 ]
Broucke, Seppe vanden [1 ,2 ]
De Weerdt, Jochen [1 ]
机构
[1] Katholieke Univ Leuven, Res Ctr Informat Syst Engn LIRIS, Leuven, Belgium
[2] Univ Ghent, Dept Business Informat & Operat Management, Ghent, Belgium
关键词
Process mining; Predictive process analytics; LSTM; Fitness; Precision; Generalization;
D O I
10.1007/s10844-022-00765-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Various methods using machine and deep learning have been proposed to tackle different tasks in predictive process monitoring, forecasting for an ongoing case e.g. the most likely next event or suffix, its remaining time, or an outcome-related variable. Recurrent neural networks (RNNs), and more specifically long short-term memory nets (LSTMs), stand out in terms of popularity. In this work, we investigate the capabilities of such an LSTM to actually learn the underlying process model structure of an event log. We introduce an evaluation framework that combines variant-based resampling and custom metrics for fitness, precision and generalization. We evaluate 4 hypotheses concerning the learning capabilities of LSTMs, the effect of overfitting countermeasures, the level of incompleteness in the training set and the level of parallelism in the underlying process model. We confirm that LSTMs can struggle to learn process model structure, even with simplistic process data and in a very lenient setup. Taking the correct anti-overfitting measures can alleviate the problem. However these measures did not present themselves to be optimal when selecting hyperparameters purely on predicting accuracy. We also found that decreasing the amount of information seen by the LSTM during training, causes a sharp drop in generalization and precision scores. In our experiments, we could not identify a relationship between the extent of parallelism in the model and the generalization capability, but they do indicate that the process' complexity might have impact.
引用
收藏
页码:27 / 51
页数:25
相关论文
共 50 条
  • [1] Can recurrent neural networks learn process model structure?
    Jari Peeperkorn
    Seppe vanden Broucke
    Jochen De Weerdt
    Journal of Intelligent Information Systems, 2023, 61 : 27 - 51
  • [2] Can Deep Neural Networks Learn Process Model Structure? An Assessment Framework and Analysis
    Peeperkorn, Jari
    vanden Broucke, Seppe
    De Weerdt, Jochen
    PROCESS MINING WORKSHOPS, ICPM 2021, 2022, 433 : 127 - 139
  • [3] USING RECURRENT NEURAL NETWORKS TO LEARN THE STRUCTURE OF INTERCONNECTION NETWORKS
    GOUDREAU, MW
    GILES, CL
    NEURAL NETWORKS, 1995, 8 (05) : 793 - 804
  • [4] Can recurrent neural networks learn natural language grammars?
    Lawrence, S
    Giles, CL
    Fong, S
    ICNN - 1996 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS. 1-4, 1996, : 1853 - 1858
  • [5] Can SGD Learn Recurrent Neural Networks with Provable Generalization?
    Allen-Zhu, Zeyuan
    Li, Yuanzhi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [6] Simple algorithm for recurrent neural networks that can learn sequence completion
    Szita, I
    Lorincz, A
    2004 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2004, : 183 - 188
  • [7] Recurrent neural networks can learn to implement symbol-sensitive counting
    Rodriguez, P
    Wiles, J
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 10, 1998, 10 : 87 - 93
  • [8] Can neural networks learn finite elements?
    Novo, Julia
    Terres, Eduardo
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2025, 453
  • [9] Learning to Learn and Compositionality with Deep Recurrent Neural Networks
    de Freitas, Nando
    KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, : 3 - 3
  • [10] Second-order recurrent neural networks can learn regular grammars from noisy strings
    Carrasco, RC
    Forcada, ML
    FROM NATURAL TO ARTIFICIAL NEURAL COMPUTATION, 1995, 930 : 605 - 610