Always be Pre-Training: Representation Learning for Network Intrusion Detection with GNNs

被引:1
|
作者
Gu, Zhengyao [1 ]
Lopez, Diego Troy [2 ]
Alrahis, Lilas [3 ]
Sinanoglu, Ozgur [3 ]
机构
[1] NYU, Ctr Data Sci, New York, NY 10012 USA
[2] NYU, Res Technol Serv, New York, NY USA
[3] New York Univ Abu Dhabi, Abu Dhabi, U Arab Emirates
关键词
Intrusion detection; machine learning; graph neural network; NIDS; few-shot learning; self-supervised learning; INTERNET; THINGS; ATTACK; IOT;
D O I
10.1109/ISQED60706.2024.10528371
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Graph neural network-based network intrusion detection systems have recently demonstrated state-of-the-art performance on benchmark datasets. Nevertheless, these methods suffer from a reliance on target encoding for data pre-processing, limiting widespread adoption due to the associated need for annotated labels-a cost-prohibitive requirement. In this work, we propose a solution involving in-context pre-training and the utilization of dense representations for categorical features to jointly overcome the label-dependency limitation. Our approach exhibits remarkable data efficiency, achieving over 98% of the performance of the supervised state-of-the-art with less than 4% labeled data on the NF-UQ-NIDS-V2 dataset.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] Robot Learning with Sensorimotor Pre-training
    Radosavovic, Ilija
    Shi, Baifeng
    Fu, Letian
    Goldberg, Ken
    Darrell, Trevor
    Malik, Jitendra
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [22] Unsupervised Pre-Training for Detection Transformers
    Dai, Zhigang
    Cai, Bolun
    Lin, Yugeng
    Chen, Junying
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (11) : 12772 - 12782
  • [23] Graph-Text Multi-Modal Pre-training for Medical Representation Learning
    Park, Sungjin
    Bae, Seongsu
    Kim, Jiho
    Kim, Tackeun
    Choi, Edward
    CONFERENCE ON HEALTH, INFERENCE, AND LEARNING, VOL 174, 2022, 174 : 261 - 281
  • [24] Deep Pre-Training Transformers for Scientific Paper Representation
    Wang, Jihong
    Yang, Zhiguang
    Cheng, Zhanglin
    ELECTRONICS, 2024, 13 (11)
  • [25] CanvasEmb: Learning Layout Representation with Large-scale Pre-training for Graphic Design
    Xie, Yuxi
    Huang, Danqing
    Wang, Jinpeng
    Lin, Chin-Yew
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 4100 - 4108
  • [26] Multi-Faceted Knowledge-Driven Pre-Training for Product Representation Learning
    Zhang, Denghui
    Liu, Yanchi
    Yuan, Zixuan
    Fu, Yanjie
    Chen, Haifeng
    Xiong, Hui
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (07) : 7239 - 7250
  • [27] UNISPEECH-SAT: UNIVERSAL SPEECH REPRESENTATION LEARNING WITH SPEAKER AWARE PRE-TRAINING
    Chen, Sanyuan
    Wu, Yu
    Wang, Chengyi
    Chen, Zhengyang
    Chen, Zhuo
    Liu, Shujie
    Wu, Jian
    Qian, Yao
    Wei, Furu
    Li, Jinyu
    Yu, Xiangzhan
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6152 - 6156
  • [28] RAPT: Pre-training of Time-Aware Transformer for Learning Robust Healthcare Representation
    Ren, Houxing
    Wang, Jingyuan
    Zhao, Wayne Xin
    Wu, Ning
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 3503 - 3511
  • [29] From Pre-Training to Meta-Learning: A Journey in Low-Resource-Language Representation Learning
    Zaikis, Dimitrios
    Vlahavas, Ioannis
    IEEE ACCESS, 2023, 11 : 115951 - 115967
  • [30] Learning to Sample Replacements for ELECTRA Pre-Training
    Hao, Yaru
    Dong, Li
    Bao, Hangbo
    Xu, Ke
    Wei, Furu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 4495 - 4506