AGCNT: Adaptive Graph Convolutional Network for Transformer-based Long Sequence Time-Series Forecasting

被引:2
|
作者
Su, Hongyang [1 ]
Wang, Xiaolong [1 ]
Qin, Yang [1 ]
机构
[1] Harbin Inst Technol, Shenzhen, Peoples R China
关键词
long sequence time-series forecasting; transformer; adaptive graph convolution; probsparse graph self-attention;
D O I
10.1145/3459637.3482054
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Long sequence time-series forecasting(LSTF) plays an important role in a variety of real-world application scenarios, such as electricity forecasting, weather forecasting, and traffic flow forecasting. It has previously been observed that transformer-based models have achieved outstanding results on LSTF tasks, which can reduce the complexity of the model and maintain stable prediction accuracy. Nevertheless, there are still some issues that limit the performance of transformer-based models for LSTF tasks: (i) the potential correlation between sequences is not considered; (ii) the inherent structure of encoder-decoder is difficult to expand after being optimized from the aspect of complexity. In order to solve these two problems, we propose a transformer-based model, named AGCNT, which is efficient and can capture the correlation between the sequences in the multivariate LSTF task without causing the memory bottleneck. Specifically, AGCNT has several characteristics: (i) a probsparse adaptive graph self-attention, which maps long sequences into a low-dimensional dense graph structure with an adaptive graph generation and captures the relationships between sequences with an adaptive graph convolution; (ii) the stacked encoder with distilling probsparse graph self-attention integrates the graph attention mechanism and retains the dominant attention of the cascade layer, which preserves the correlation between sparse queries from long sequences; (iii) the stacked decoder with generative inference generates all prediction values in one forward operation, which can improve the inference speed of long-term predictions. Experimental results on 4 large-scale datasets demonstrate the AGCNT outperforms state-of-the-art baselines.
引用
收藏
页码:3439 / 3442
页数:4
相关论文
共 50 条
  • [1] RSMformer: an efficient multiscale transformer-based framework for long sequence time-series forecasting
    Tong, Guoxiang
    Ge, Zhaoyuan
    Peng, Dunlu
    APPLIED INTELLIGENCE, 2024, 54 (02) : 1275 - 1296
  • [2] RSMformer: an efficient multiscale transformer-based framework for long sequence time-series forecasting
    Guoxiang Tong
    Zhaoyuan Ge
    Dunlu Peng
    Applied Intelligence, 2024, 54 (2) : 1275 - 1296
  • [3] ST_AGCNT: Traffic Speed Forecasting Based on Spatial-Temporal Adaptive Graph Convolutional Network with Transformer
    Cheng, Rongjun
    Liu, Mengxia
    Xu, Yuanzi
    SUSTAINABILITY, 2025, 17 (05)
  • [4] Graphformer: Adaptive graph correlation transformer for multivariate long sequence time series forecasting
    Wang, Yijie
    Long, Hao
    Zheng, Linjiang
    Shang, Jiaxing
    KNOWLEDGE-BASED SYSTEMS, 2024, 285
  • [5] TIformer: A Transformer-Based Framework for Time-Series Forecasting with Missing Data
    Ding, Zuocheng
    Chen, Yufan
    Wang, Hanchen
    Wang, Xiaoyang
    Zhang, Wenjie
    Zhang, Ying
    DATABASES THEORY AND APPLICATIONS, ADC 2024, 2025, 15449 : 71 - 84
  • [6] Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting
    Zhou, Haoyi
    Zhang, Shanghang
    Peng, Jieqi
    Zhang, Shuai
    Li, Jianxin
    Xiong, Hui
    Zhang, Wancai
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 11106 - 11115
  • [7] Enhancing Transformer-based models for long sequence time series forecasting via structured matrix
    Zhang, Zhicheng
    Wang, Yong
    Tan, Shaoqi
    Xia, Bowei
    Luo, Yujie
    NEUROCOMPUTING, 2025, 625
  • [8] Transformer-Based Graph Convolutional Network for Sentiment Analysis
    AlBadani, Barakat
    Shi, Ronghua
    Dong, Jian
    Al-Sabri, Raeed
    Moctard, Oloulade Babatounde
    APPLIED SCIENCES-BASEL, 2022, 12 (03):
  • [9] Multivariate long sequence time-series forecasting using dynamic graph learning
    Wang X.
    Wang Y.
    Peng J.
    Zhang Z.
    Journal of Ambient Intelligence and Humanized Computing, 2023, 14 (06) : 7679 - 7693
  • [10] Spatiotemporal Transformer Neural Network for Time-Series Forecasting
    You, Yujie
    Zhang, Le
    Tao, Peng
    Liu, Suran
    Chen, Luonan
    ENTROPY, 2022, 24 (11)