A malicious network traffic detection model based on bidirectional temporal convolutional network with multi-head self-attention mechanism

被引:8
|
作者
Cai, Saihua [1 ,2 ]
Xu, Han [1 ]
Liu, Mingjie [1 ]
Chen, Zhilin [1 ]
Zhang, Guofeng [3 ]
机构
[1] Jiangsu Univ, Sch Comp Sci & Commun Engn, Zhenjiang 212013, Peoples R China
[2] Jiangsu Univ, Jiangsu Key Lab Secur Technol Ind Cyberspace, Zhenjiang 212013, Peoples R China
[3] Taishan Univ, Sch Informat Sci & Technol, Tai An 271000, Peoples R China
基金
中国国家自然科学基金;
关键词
Malicious network traffic detection; Bidirectional temporal convolutional network; Multi -head self -attention mechanism; Cross -entropy loss function; Deep learning;
D O I
10.1016/j.cose.2023.103580
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The increasingly frequent network intrusions have brought serious impacts to the production and life, thus malicious network traffic detection has received more and more attention in recent years. However, the traditional rule matching-based and machine learning-based malicious network traffic detection methods have the problems of relying on human experience as well as low detection efficiency. The continuous development of deep learning technology provides new ideas to solve malicious network traffic detection, and the deep learning models are also widely used in the field of malicious network traffic detection. Compared with other deep learning models, bidirectional temporal convolutional network (BiTCN) has achieved better detection results due to its ability to obtain bidirectional semantic features of network traffic, but it does not consider the different meanings as well as different importance of different subsequence segments in network traffic sequences; In addition, the loss function used in BiTCN is the negative log likelihood function, which may lead to overfitting problems when facing multi-classification problems and data imbalance problems. To solve these problems, this paper proposes a malicious network traffic detection model based on BiTCN and multi-head self-attention (MHSA) mechanism, namely BiTCN_MHSA, it innovatively uses the MHSA mechanism to assign different weights to different subsequences of network traffic, thus making the model more focused on the characteristics of malicious network traffic as well as improving the efficiency of processing global network traffic; Moreover, it also changes its loss function to a cross-entropy loss function to penalize misclassification more severely, thereby speeding up the convergence. Finally, extensive experiments are conduced to evaluate the efficiency of proposed BiTCN_MHSA model on two public network traffic, the experimental results verify that the proposed BiTCN_MHSA model outperforms six state-of-the-arts in precision, recall, F1-measure and accuracy.
引用
收藏
页数:17
相关论文
共 50 条
  • [41] Multi-Head Self-Attention Model for Classification of Temporal Lobe Epilepsy Subtypes
    Gu, Peipei
    Wu, Ting
    Zou, Mingyang
    Pan, Yijie
    Guo, Jiayang
    Xiahou, Jianbing
    Peng, Xueping
    Li, Hailong
    Ma, Junxia
    Zhang, Ling
    FRONTIERS IN PHYSIOLOGY, 2020, 11
  • [42] MSnet: Multi-Head Self-Attention Network for Distantly Supervised Relation Extraction
    Sun, Tingting
    Zhang, Chunhong
    Ji, Yang
    Hu, Zheng
    IEEE ACCESS, 2019, 7 : 54472 - 54482
  • [43] A Supervised Multi-Head Self-Attention Network for Nested Named Entity Recognition
    Xu, Yongxiu
    Huang, Heyan
    Feng, Chong
    Hu, Yue
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 14185 - 14193
  • [44] SQL Injection Detection Based on Lightweight Multi-Head Self-Attention
    Lo, Rui-Teng
    Hwang, Wen-Jyi
    Tai, Tsung-Ming
    APPLIED SCIENCES-BASEL, 2025, 15 (02):
  • [45] DILATED RESIDUAL NETWORK WITH MULTI-HEAD SELF-ATTENTION FOR SPEECH EMOTION RECOGNITION
    Li, Runnan
    Wu, Zhiyong
    Jia, Jia
    Zhao, Sheng
    Meng, Helen
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 6675 - 6679
  • [46] A novel two-stream multi-head self-attention convolutional neural network for bearing fault diagnosis
    Ren, Hang
    Liu, Shaogang
    Wei, Fengmei
    Qiu, Bo
    Zhao, Dan
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART C-JOURNAL OF MECHANICAL ENGINEERING SCIENCE, 2024, 238 (11) : 5393 - 5405
  • [47] MS-Pointer Network: Abstractive Text Summary Based on Multi-Head Self-Attention
    Guo, Qian
    Huang, Jifeng
    Xiong, Naixue
    Wang, Pan
    IEEE ACCESS, 2019, 7 : 138603 - 138613
  • [48] Convolutional multi-head self-attention on memory for aspect sentiment classification
    Zhang, Yaojie
    Xu, Bing
    Zhao, Tiejun
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2020, 7 (04) : 1038 - 1044
  • [49] Multi head self-attention gated graph convolutional network based multi-attack intrusion detection in MANET
    Reka, R.
    Karthick, R.
    Ram, R. Saravana
    Singh, Gurkirpal
    COMPUTERS & SECURITY, 2024, 136
  • [50] Convolutional Multi-Head Self-Attention on Memory for Aspect Sentiment Classification
    Yaojie Zhang
    Bing Xu
    Tiejun Zhao
    IEEE/CAA Journal of Automatica Sinica, 2020, 7 (04) : 1038 - 1044