Recurrence and Self-attention vs the Transformer for Time-Series Classification: A Comparative Study

被引:7
|
作者
Katrompas, Alexander [1 ]
Ntakouris, Theodoros [2 ]
Metsis, Vangelis [1 ]
机构
[1] Texas State Univ, San Marcos, TX 78666 USA
[2] Univ Patras, Patras, Greece
关键词
D O I
10.1007/978-3-031-09342-5_10
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently the transformer has established itself as the state-of-the-art in text processing and has demonstrated impressive results in image processing, leading to the decline in the use of recurrence in neural network models. As established in the seminal paper, Attention Is All You Need, recurrence can be removed in favor of a simpler model using only self-attention. While transformers have shown themselves to be robust in a variety of text and image processing tasks, these tasks all have one thing in common; they are inherently non-temporal. Although transformers are also finding success in modeling time-series data, they also have their limitations as compared to recurrent models. We explore a class of problems involving classification and prediction from time-series data and show that recurrence combined with self-attention can meet or exceed the transformer architecture performance. This particular class of problem, temporal classification, and prediction of labels through time from time-series data is of particular importance to medical data sets which are often time-series based (Source code: https://github.com/imics-lab/recurrence-with-self-attention).
引用
收藏
页码:99 / 109
页数:11
相关论文
共 50 条
  • [1] Self-attention for raw optical Satellite Time Series Classification
    Russwurm, Marc
    Koerner, Marco
    [J]. ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2020, 169 : 421 - 435
  • [2] STING: Self-attention based Time-series Imputation Networks using GAN
    Oh, Eunkyu
    Kim, Taehun
    Ji, Yunhu
    Khyalia, Sushil
    [J]. 2021 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2021), 2021, : 1264 - 1269
  • [3] Sparse self-attention guided generative adversarial networks for time-series generation
    Nourhan Ahmed
    Lars Schmidt-Thieme
    [J]. International Journal of Data Science and Analytics, 2023, 16 : 421 - 434
  • [4] Sparse self-attention guided generative adversarial networks for time-series generation
    Ahmed, Nourhan
    Schmidt-Thieme, Lars
    [J]. INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2023, 16 (04) : 421 - 434
  • [5] Multiscale echo self-attention memory network for multivariate time series classification
    Lyu, Huizi
    Huang, Desen
    Li, Sen
    Ma, Qianli
    Ng, Wing W. Y.
    [J]. NEUROCOMPUTING, 2023, 520 : 60 - 72
  • [6] Enformer: Encoder-Based Sparse Periodic Self-Attention Time-Series Forecasting
    Wang, Na
    Zhao, Xianglian
    [J]. IEEE ACCESS, 2023, 11 : 112004 - 112014
  • [7] Rethink the Top-u Attention in Sparse Self-attention for Long Sequence Time-Series Forecasting
    Meng, Xiangxu
    Li, Wei
    Gaber, Tarek
    Zhao, Zheng
    Chen, Chuhao
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VI, 2023, 14259 : 256 - 267
  • [8] Attention Augmented Convolutional Transformer for Tabular Time-series
    Shankaranarayana, Sharath M.
    Runje, Davor
    [J]. 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS ICDMW 2021, 2021, : 537 - 541
  • [9] Remote Sensing Time Series Classification Based on Self-Attention Mechanism and Time Sequence Enhancement
    Liu, Jingwei
    Yan, Jining
    Wang, Lizhe
    Huang, Liang
    He, Haixu
    Liu, Hong
    [J]. REMOTE SENSING, 2021, 13 (09)
  • [10] An improved self-attention for long-sequence time-series data forecasting with missing values
    Zhang, Zhi-cheng
    Wang, Yong
    Peng, Jian-jian
    Duan, Jun-ting
    [J]. NEURAL COMPUTING & APPLICATIONS, 2024, 36 (08): : 3921 - 3940