STRec: Sparse Transformer for Sequential Recommendations

被引:9
|
作者
Li, Chengxi [1 ]
Wang, Yejing [1 ]
Liu, Qidong [1 ,2 ]
Zhao, Xiangyu [1 ]
Wang, Wanyu [1 ]
Wang, Yiqi [3 ]
Zou, Lixin [4 ]
Fan, Wenqi [5 ]
Li, Qing [5 ]
机构
[1] City Univ Hong Kong, Hong Kong, Peoples R China
[2] Xi An Jiao Tong Univ, Xian, Peoples R China
[3] Michigan State Univ, E Lansing, MI 48824 USA
[4] Wuhan Univ, Wuhan, Peoples R China
[5] Hong Kong Polytech Univ, Hong Kong, Peoples R China
来源
PROCEEDINGS OF THE 17TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2023 | 2023年
关键词
recommendation system; sequential recommendation; efficient transformer;
D O I
10.1145/3604915.3608779
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid evolution of transformer architectures, researchers are exploring their application in sequential recommender systems (SRSs) and presenting promising performance on SRS tasks compared with former SRS models. However, most existing transformer-based SRS frameworks retain the vanilla attention mechanism, which calculates the attention scores between all item-item pairs. With this setting, redundant item interactions can harm the model performance and consume much computation time and memory. In this paper, we identify the sparse attention phenomenon in transformer-based SRS models and propose Sparse Transformer for sequential Recommendation tasks (STRec) to achieve the efficient computation and improved performance. Specifically, we replace self-attention with cross-attention, making the model concentrate on the most relevant item interactions. To determine these necessary interactions, we design a novel sampling strategy to detect relevant items based on temporal information. Extensive experimental results validate the effectiveness of STRec, which achieves the state-of-the-art accuracy while reducing 54% inference time and 70% memory cost. We also provide massive extended experiments to further investigate the property of our framework.
引用
收藏
页码:101 / 111
页数:11
相关论文
共 50 条
  • [1] MAE4Rec: Storage-saving Transformer for Sequential Recommendations
    Zhao, Kesen
    Zhao, Xiangyu
    Zhang, Zijian
    Li, Muyang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 2681 - 2690
  • [2] Attentive Flexible Translation Embedding in Top-N Sparse Sequential Recommendations
    Seo, Min-Ji
    Kim, Myung-Ho
    INTERNATIONAL JOURNAL OF INTERACTIVE MULTIMEDIA AND ARTIFICIAL INTELLIGENCE, 2023, 8 (04): : 56 - 66
  • [3] Dual Contrastive Learning and Dual Bi-directional Transformer Encoders for Sequential Recommendations
    Wang, Li-e
    Chang, Hengtong
    Wei, Rongwen
    Li, Xianxian
    Sun, Zhigang
    Li, Yongdong
    Wei, Yi
    Meng, LingHui
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 1388 - 1393
  • [4] Improving Transformer-based Sequential Conversational Recommendations through Knowledge Graph Embeddings
    Petruzzelli, Alessandro
    Martina, Alessandro Francesco Maria
    Spillo, Giuseppe
    Musto, Cataldo
    de Gemmis, Marco
    Lops, Pasquale
    Semeraro, Giovanni
    PROCEEDINGS OF THE 32ND ACM CONFERENCE ON USER MODELING, ADAPTATION AND PERSONALIZATION, UMAP 2024, 2024, : 172 - 182
  • [5] Sparse Universal Transformer
    Tan, Shawn
    Shen, Yikang
    Chen, Zhenfang
    Courville, Aaron
    Gan, Chuang
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 169 - 179
  • [6] SparseSwin: Swin transformer with sparse transformer block
    Pinasthika, Krisna
    Laksono, Blessius Sheldo Putra
    Irsal, Riyandi Banovbi Putera
    Shabiyya, Syifa Hukma
    Yudistira, Novanto
    NEUROCOMPUTING, 2024, 580
  • [7] Sequential Sparse Matching Pursuit
    Berinde, Radu
    Indyk, Piotr
    2009 47TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING, VOLS 1 AND 2, 2009, : 36 - +
  • [8] Sequential Testing for Sparse Recovery
    Malloy, Matthew L.
    Nowak, Robert D.
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2014, 60 (12) : 7862 - 7873
  • [9] Layer Sparse Transformer for Speech Recognition
    Wang, Peng
    Guo, Zhiyuan
    Xie, Fei
    2023 IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH, ICKG, 2023, : 269 - 273
  • [10] Sparse Graph Transformer With Contrastive Learning
    Zhang, Chun-Yang
    Fang, Wu-Peng
    Cai, Hai-Chun
    Chen, C. L. Philip
    Lin, Yue-Na
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (01): : 892 - 904