EGAD: Evolving Graph Representation Learning with Self-Attention and Knowledge Distillation for Live Video Streaming Events

被引:3
|
作者
Antaris, Stefanos [1 ,2 ]
Rafailidis, Dimitrios [3 ]
Girdzijauskas, Sarunas [1 ]
机构
[1] KTH Royal Inst Technol, Stockholm, Sweden
[2] HiveStreaming AB, Stockholm, Sweden
[3] Maastricht Univ, Maastricht, Netherlands
关键词
Graph representation learning; live video streaming; evolving graphs; knowledge distillation;
D O I
10.1109/BigData50022.2020.9378219
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this study, we present a dynamic graph representation learning model on weighted graphs to accurately predict the network capacity of connections between viewers in a live video streaming event. We propose EGAD, a neural network architecture to capture the graph evolution by introducing a self-attention mechanism on the weights between consecutive graph convolutional networks. In addition, we account for the fact that neural architectures require a huge amount of parameters to train, thus increasing the online inference latency and negatively influencing the user experience in a live video streaming event. To address the problem of the high online inference of a vast number of parameters, we propose a knowledge distillation strategy. In particular, we design a distillation loss function, aiming to first pretrain a teacher model on offline data, and then transfer the knowledge from the teacher to a smaller student model with less parameters. We evaluate our proposed model on the link prediction task on three real-world datasets, generated by live video streaming events. The events lasted 80 minutes and each viewer exploited the distribution solution provided by the company Hive Streaming AB. The experiments demonstrate the effectiveness of the proposed model in terms of link prediction accuracy and number of required parameters, when evaluated against state-of-the-art approaches. In addition, we study the distillation performance of the proposed model in terms of compression ratio for different distillation strategies, where we show that the proposed model can achieve a compression ratio up to 15:100, preserving high link prediction accuracy. For reproduction purposes, our evaluation datasets and implementation are publicly available at https://stefanosantaris.github.io/EGAD.
引用
收藏
页码:1455 / 1464
页数:10
相关论文
共 50 条
  • [31] Self-attention empowered graph convolutional network for structure learning and node embedding
    Jiang, Mengying
    Liu, Guizhong
    Su, Yuanchao
    Wu, Xinliang
    PATTERN RECOGNITION, 2024, 153
  • [32] SAM-Net: Self-Attention based Feature Matching with Spatial Transformers and Knowledge Distillation
    Kelenyi, Benjamin
    Domsa, Victor
    Tamas, Levente
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 242
  • [33] Self-attention presents low-dimensional knowledge graph embeddings for link prediction
    Baghershahi, Peyman
    Hosseini, Reshad
    Moradi, Hadi
    KNOWLEDGE-BASED SYSTEMS, 2023, 260
  • [34] Class token and knowledge distillation for multi-head self-attention speaker verification systems
    Mingote, Victoria
    Miguel, Antonio
    Ortega, Alfonso
    Lleida, Eduardo
    DIGITAL SIGNAL PROCESSING, 2023, 133
  • [35] Towards accurate diagnosis: exploring knowledge distillation and self-attention in multimodal medical image fusion
    Radhika, P.
    Bobby, J. Sofia
    Francis, Sheeja V.
    Femina, M. A.
    JOURNAL OF EXPERIMENTAL & THEORETICAL ARTIFICIAL INTELLIGENCE, 2024,
  • [36] Multi-task Learning for User Engagement and Adoption in Live Video Streaming Events
    Antaris, Stefanos
    Rafailidis, Dimitrios
    Arriaza, Romina
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: APPLIED DATA SCIENCE TRACK, PT V, 2021, 12979 : 463 - 478
  • [37] SAGSleepNet: A deep learning model for sleep staging based on self-attention graph of polysomnography
    Jin, Zheng
    Jia, Kebin
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 86
  • [38] Cross-Modal Graph Knowledge Representation and Distillation Learning for Land Cover Classification
    Wang, Wenzhen
    Liu, Fang
    Liao, Wenzhi
    Xiao, Liang
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [39] Self-attention Multi-view Representation Learning with Diversity-promoting Complementarity
    Liu, Jian-wei
    Ding, Xi-hao
    Lu, Run-kun
    Luo, Xionglin
    PROCEEDINGS OF THE 32ND 2020 CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2020), 2020, : 3972 - 3978
  • [40] Attribute Network Representation Learning Based on Generative Adversarial Network and Self-attention Mechanism
    Li, Shanshan
    Tang, Meiling
    Dong, Yingnan
    International Journal of Network Security, 2024, 26 (01) : 51 - 58