Detecting multi-stage attacks using sequence-to-sequence model

被引:13
|
作者
Zhou, Peng [1 ]
Zhou, Gongyan [1 ]
Wu, Dakui [1 ]
Fei, Minrui [1 ]
机构
[1] Shanghai Univ, Shanghai Key Lab Power Stn Automat Technol, Shanghai, Peoples R China
基金
上海市自然科学基金; 中国国家自然科学基金;
关键词
Multi-stage attack; Intrusion detection; Sequence-to-sequence model; Encoder-decoder architecture; Long-short term memory (LSTM) network;
D O I
10.1016/j.cose.2021.102203
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-stage attack is a kind of sophisticated intrusion strategy that has been widely used for penetrating the well protected network infrastructures. To detect such attacks, state-of-theart research advocates the use of hidden markov model (HMM). However, despite the HMM can model the relationships and dependencies among different alerts and stages for detection, they cannot handle well the stage dependencies buried in a longer sequence of alerts. In this paper, we tackle the challenge of the stages' long-term dependency and propose a new detection solution using a sequence-to-sequence (seq2seq) model. The basic idea is to encode a sequence of alerts (i.e., detector's observation) into a latent feature vector using a long-short term memory (LSTM) network and then decode this vector to a sequence of predicted attacking stages with another LSTM. By the encoder-decoder collaboration, we can decouple the local constraint between the observed alerts and the potential attacking stages, and thus able to take the full knowledge of all the alerts for the detection of stages in a sequence basis. By the LSTM, we can learn to "forget" irrelevant alerts and thereby have more opportunities to "remember" the long-term dependency between different stages for our sequence detection. To evaluate our model's effectiveness, we have conducted extensive experiments using four public datasets, all of which include simulated or re-constructed samples of real-world multi-stage attacks in controlled testbeds. Our results have successfully confirmed the better detection performance of our model compared with the previous HMM solutions. (c) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Sequence-to-Sequence Model with Attention for Time Series Classification
    Tang, Yujin
    Xu, Jianfeng
    Matsumoto, Kazunori
    Ono, Chihiro
    2016 IEEE 16TH INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW), 2016, : 503 - 510
  • [22] MITIGATING THE IMPACT OF SPEECH RECOGNITION ERRORS ON CHATBOT USING SEQUENCE-TO-SEQUENCE MODEL
    Chen, Pin-Jung
    Hsu, I-Hung
    Huang, Yi-Yao
    Lee, Hung-Yi
    2017 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2017, : 497 - 503
  • [23] A Hierarchical Sequence-to-Sequence Model for Korean POS Tagging
    Jin, Guozhe
    Yu, Zhezhou
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2021, 20 (02)
  • [24] Myanmar News Headline Generation with Sequence-to-Sequence model
    Thu, Yamin
    Pa, Win Pa
    PROCEEDINGS OF 2020 23RD CONFERENCE OF THE ORIENTAL COCOSDA INTERNATIONAL COMMITTEE FOR THE CO-ORDINATION AND STANDARDISATION OF SPEECH DATABASES AND ASSESSMENT TECHNIQUES (ORIENTAL-COCOSDA 2020), 2020, : 117 - 122
  • [25] Sequence-to-Sequence model for Building Energy Consumption Prediction
    Kim, Marie
    Jun, JongAm
    Kim, Nasoo
    Song, YuJin
    Pyo, Cheol Sik
    2018 INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGY CONVERGENCE (ICTC), 2018, : 1243 - 1245
  • [26] BARThez: a Skilled Pretrained French Sequence-to-Sequence Model
    Eddine, Moussa Kamal
    Tixier, Antoine J-P
    Vazirgiannis, Michalis
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 9369 - 9390
  • [27] SAMAF: Sequence-to-sequence Autoencoder Model for Audio Fingerprinting
    Baez-Suarez, Abraham
    Shah, Nolan
    Nolazco-Flores, Juan Arturo
    Huang, Shou-Hsuan S.
    Gnawali, Omprakash
    Shi, Weidong
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2020, 16 (02)
  • [28] Attention Strategies for Multi-Source Sequence-to-Sequence Learning
    Libovicky, Jindrich
    Helcl, Jindrich
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 2, 2017, : 196 - 202
  • [29] Sequence-to-Sequence Multi-Modal Speech In-Painting
    Elyaderani, Mahsa Kadkhodaei
    Shirani, Shahram
    INTERSPEECH 2023, 2023, : 829 - 833
  • [30] Rainfall-runoff modeling using LSTM-based multi-state-vector sequence-to-sequence model
    Yin, Hanlin
    Zhang, Xiuwei
    Wang, Fandu
    Zhang, Yanning
    Xia, Runliang
    Jin, Jin
    JOURNAL OF HYDROLOGY, 2021, 598