Sequential Recommendation via Stochastic Self-Attention

被引:59
|
作者
Fan, Ziwei [1 ,5 ]
Liu, Zhiwei [1 ]
Wang, Yu [1 ]
Wang, Alice [2 ]
Nazari, Zahra [2 ]
Zheng, Lei [3 ]
Peng, Hao [4 ]
Yu, Philip S. [1 ]
机构
[1] Univ Illinois, Dept Comp Sci, Chicago, IL 60680 USA
[2] Spotify, New York, NY USA
[3] Pinterest Inc, Chicago, IL USA
[4] Beihang Univ, Sch Cyber Sci & Technol, Beijing, Peoples R China
[5] Spotify Res, New York, NY USA
关键词
Sequential Recommendation; Transformer; Self-Attention; Uncertainty;
D O I
10.1145/3485447.3512077
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Sequential recommendation models the dynamics of a user's previous behaviors in order to forecast the next item, and has drawn a lot of attention. Transformer-based approaches, which embed items as vectors and use dot-product self-attention to measure the relationship between items, demonstrate superior capabilities among existing sequential methods. However, users' real-world sequential behaviors are uncertain rather than deterministic, posing a significant challenge to present techniques. We further suggest that dot-product-based approaches cannot fully capture collaborative transitivity, which can be derived in item-item transitions inside sequences and is beneficial for cold start items. We further argue that BPR loss has no constraint on positive and sampled negative items, which misleads the optimization. We propose a novel STOchastic Self-Attention (STOSA) to overcome these issues. STOSA, in particular, embeds each item as a stochastic Gaussian distribution, the covariance of which encodes the uncertainty. We devise a novel Wasserstein Self-Attention module to characterize item-item position-wise relationships in sequences, which effectively incorporates uncertainty into model training. Wasserstein attentions also enlighten the collaborative transitivity learning as it satisfies triangle inequality. Moreover, we introduce a novel regularization term to the ranking loss, which assures the dissimilarity between positive and the negative items. Extensive experiments on five real-world benchmark datasets demonstrate the superiority of the proposed model over state-of-the-art baselines, especially on cold start items. The code is available in https://github.com/zfan20/STOSA.
引用
收藏
页码:2036 / 2047
页数:12
相关论文
共 50 条
  • [41] Self-attention Based Collaborative Neural Network for Recommendation
    Ma, Shengchao
    Zhu, Jinghua
    WIRELESS ALGORITHMS, SYSTEMS, AND APPLICATIONS, WASA 2019, 2019, 11604 : 235 - 246
  • [42] Hashtag Recommendation Using LSTM Networks with Self-Attention
    Shen, Yatian
    Li, Yan
    Sun, Jun
    Ding, Wenke
    Shi, Xianjin
    Zhang, Lei
    Shen, Xiajiong
    He, Jing
    CMC-COMPUTERS MATERIALS & CONTINUA, 2019, 61 (03): : 1261 - 1269
  • [43] Exception Handling Recommendation Based on Self-Attention Network
    Lin, Kai
    Tao, Chuanqi
    Huang, Zhiqiu
    2021 IEEE INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING WORKSHOPS (ISSREW 2021), 2021, : 282 - 283
  • [44] Sequential Recommendation Based on Long-Term and Short-Term User Behavior with Self-attention
    Wei, Xing
    Zuo, Xianglin
    Yang, Bo
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2019, PT I, 2019, 11775 : 72 - 83
  • [45] An Improved Sequential Recommendation Algorithm based on Short-Sequence Enhancement and Temporal Self-Attention Mechanism
    Ni, Jianjun
    Tang, Guangyi
    Shen, Tong
    Cai, Yu
    Cao, Weidong
    COMPLEXITY, 2022, 2022
  • [46] Collaborative Self-Attention Network for Session-based Recommendation
    Luo, Anjing
    Zhao, Pengpeng
    Liu, Yanchi
    Zhuang, Fuzhen
    Wang, Deqing
    Xu, Jiajie
    Fang, Junhua
    Sheng, Victor S.
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 2591 - 2597
  • [47] Context-embedded hypergraph attention network and self-attention for session recommendation
    Zhang, Zhigao
    Zhang, Hongmei
    Zhang, Zhifeng
    Wang, Bin
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [48] A Service Recommendation Algorithm Based on Self-Attention Mechanism and DeepFM
    Deng, Li Ping
    Guo, Bing
    Zheng, Wen
    INTERNATIONAL JOURNAL OF WEB SERVICES RESEARCH, 2023, 20 (01)
  • [49] A collaborative filtering recommendation algorithm based on DeepWalk and self-attention
    Guo, Jiaming
    Wen, Hong
    Huang, Weihong
    Yang, Ce
    INTERNATIONAL JOURNAL OF COMPUTATIONAL SCIENCE AND ENGINEERING, 2023, 26 (03) : 296 - 304
  • [50] HARSAM: A Hybrid Model for Recommendation Supported by Self-Attention Mechanism
    Peng, Dunlu
    Yuan, Weiwei
    Liu, Cong
    IEEE ACCESS, 2019, 7 : 12620 - 12629