Combining Multi-Head Attention and Sparse Multi-Head Attention Networks for Session-Based Recommendation

被引:1
|
作者
Zhao, Zhiwei [1 ]
Wang, Xiaoye [1 ]
Xiao, Yingyuan [1 ]
机构
[1] Tianjin Univ Technol, Sch Comp Sci & Engn, Tianjin, Peoples R China
关键词
session-based recommendation; multi-head attention; sparse multi-head attention;
D O I
10.1109/IJCNN54540.2023.10191924
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The core of session-based recommendation is to predict the next interactive item based on a set of anonymous user temporal or specified behavior sequences (e.g., click, browse or purchase item sequence), which is a key task of many online services today. Recently, self-attention networks have achieved remarkable success in the task of session-based recommendation. However, in session-based recommendation, some items may be clicked by mistake, and most of the current attention mechanisms assign weights to these items, resulting in the disadvantage of distraction. Although sparse attention networks can address the aforementioned issues, solely relying on sparse attention may in turn reduce the weight of some real-intent clicked items. Therefore, this paper proposes a model that combines multi-headed attention network and sparse multi-headed attention network, referred to as CMAN, which solves the drawback of assigning weights to items clicked by mistake in the traditional attention mechanism. And also prevents the drawback of reducing the weights of items that are truly clicked by some users brought by using sparse attention mechanism alone to some extent. Experiments on two real datasets show that the model outperforms some state-of-the-art models.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] Personalized multi-head self-attention network for news recommendation
    Zheng, Cong
    Song, Yixuan
    [J]. Neural Networks, 2025, 181
  • [22] MAFD: A Federated Distillation Approach with Multi-head Attention for Recommendation Tasks
    Wu, Aming
    Kwon, Young-Woo
    [J]. 38TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2023, 2023, : 1221 - 1224
  • [23] Learning Sentences Similarity By Multi-Head Attention
    Wang, Ming Yang
    Li, Chen Jiang
    Sun, Jian Dong
    Xu, Wei Ran
    Gao, Sheng
    Zhang, Ya Hao
    Wang, Pu
    Li, Jun Liang
    [J]. PROCEEDINGS OF 2018 INTERNATIONAL CONFERENCE ON NETWORK INFRASTRUCTURE AND DIGITAL CONTENT (IEEE IC-NIDC), 2018, : 16 - 19
  • [24] DOUBLE MULTI-HEAD ATTENTION FOR SPEAKER VERIFICATION
    India, Miquel
    Safari, Pooyan
    Hernando, Javier
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6144 - 6148
  • [25] VIDEO SUMMARIZATION WITH ANCHORS AND MULTI-HEAD ATTENTION
    Sung, Yi-Lin
    Hong, Cheng-Yao
    Hsu, Yen-Chi
    Liu, Tyng-Luh
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 2396 - 2400
  • [26] Financial Volatility Forecasting: A Sparse Multi-Head Attention Neural Network
    Lin, Hualing
    Sun, Qiubi
    [J]. INFORMATION, 2021, 12 (10)
  • [27] Classification of Heads in Multi-head Attention Mechanisms
    Huang, Feihu
    Jiang, Min
    Liu, Fang
    Xu, Dian
    Fan, Zimeng
    Wang, Yonghao
    [J]. KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2022, PT III, 2022, 13370 : 681 - 692
  • [28] Sequential Recommendation Using Deep Reinforcement Learning and Multi-Head Attention
    Sultan, Raneem
    Abu-Elkheir, Mervat
    [J]. 2022 56TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2022, : 258 - 262
  • [29] Software and Hardware Fusion Multi-Head Attention
    Hu, Wei
    Xu, Dian
    Liu, Fang
    Fan, Zimeng
    [J]. KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2022, PT III, 2022, 13370 : 644 - 655
  • [30] Finding the Pillars of Strength for Multi-Head Attention
    Ni, Jinjie
    Mao, Rui
    Yang, Zonglin
    Lei, Han
    Cambria, Erik
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14526 - 14540