Towards more effective encoders in pre-training for sequential recommendation

被引:1
|
作者
Sun, Ke [1 ]
Qian, Tieyun [1 ]
Zhong, Ming [1 ]
Li, Xuhui [2 ]
机构
[1] Wuhan Univ, Sch Comp Sci, Wuhan, Peoples R China
[2] Wuhan Univ, Sch Informat Management, Wuhan, Peoples R China
来源
WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS | 2023年 / 26卷 / 05期
基金
中国国家自然科学基金;
关键词
Sequential recommendation; Self-supervised learning; Pre-training; Encoder; CONTEXT;
D O I
10.1007/s11280-023-01163-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Pre-training emerges as a new learning paradigm in natural language processing and computer vision. It has also been introduced into sequential recommendation in several seminal studies for alleviating data sparsity issue. However, existing methods adopt the bidirectional transformer as the encoder which suffers from two drawbacks. One is insufficient intention modeling since the transformer architecture is suitable for extracting distributed consumption intention but cannot well catch users' concentrated and occasion consumption intentions. The other is information leakage caused by foreseeing the future item in advance during the bidirectional encoding process. To address these problems, we propose to construct more effective encoders in pre-training for sequential recommendation. Specifically, we first decouple the original bidirectional process in transformer structure into two unidirectional processes which can avoid the information leakage problem and capture the distributed consumption intention. We then employ the locality-aware convolutional neural networks (CNNs) with narrow receptive field for concentrated consumption modeling. We also introduce a random shuffle strategy to empower CNN with the ability of modeling the occasion consumption. Experiments on five datasets demonstrate that our method improves the performance of various types of downstream sequential recommendation models to a large extent, and it also generates the overall better performance than the state-of-the-art self-supervised pre-training methods.
引用
收藏
页码:2801 / 2832
页数:32
相关论文
共 50 条
  • [41] Bootstrapping ViTs: Towards Liberating Vision Transformers from Pre-training
    Zhang, Haofei
    Duan, Jiarui
    Xue, Mengqi
    Song, Jie
    Sun, Li
    Song, Mingli
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 8934 - 8943
  • [42] Learning Transferable User Representations with Sequential Behaviors via Contrastive Pre-training
    Cheng, Mingyue
    Yuan, Fajie
    Liu, Qi
    Xin, Xin
    Chen, Enhong
    2021 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2021), 2021, : 51 - 60
  • [43] Multi-stage Pre-training over Simplified Multimodal Pre-training Models
    Liu, Tongtong
    Feng, Fangxiang
    Wang, Xiaojie
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 2556 - 2565
  • [44] Improving Sequential Recommendations via Bidirectional Temporal Data Augmentation With Pre-Training
    Jiang, Juyong
    Zhang, Peiyan
    Luo, Yingtao
    Li, Chaozhuo
    Kim, Jae Boum
    Zhang, Kai
    Wang, Senzhang
    Kim, Sunghun
    Yu, Philip S.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2025, 37 (05) : 2652 - 2664
  • [45] Towards Adversarial Attack on Vision-Language Pre-training Models
    Zhang, Jiaming
    Yi, Qi
    Sang, Jitao
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 5005 - 5013
  • [46] Towards Masked Autoencoding Pre-training for Wide Area Motion Imagery
    Goley, Steve
    Pradhan, Rohan
    Welch, Austin
    GEOSPATIAL INFORMATICS XIII, 2023, 12525
  • [47] KEEP: An Industrial Pre-Training Framework for Online Recommendation via Knowledge Extraction and Plugging
    Zhang, Yujing
    Chan, Zhangming
    Xu, Shuhao
    Bian, Weijie
    Han, Shuguang
    Deng, Hongbo
    Zheng, Bo
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 3684 - 3693
  • [48] Mixed-Order Heterogeneous Graph Pre-training for Cold-Start Recommendation
    Sui, Wenzheng
    Jiang, Xiaoxia
    Ge, Weiyi
    Hu, Wei
    WEB AND BIG DATA, PT III, APWEB-WAIM 2022, 2023, 13423 : 182 - 190
  • [49] A Multi-strategy-based Pre-training Method for Cold-start Recommendation
    Hao, Bowen
    Yin, Hongzhi
    Zhang, Jing
    Li, Cuiping
    Chen, Hong
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2023, 41 (02)
  • [50] Masked Modeling Duo: Towards a Universal Audio Pre-Training Framework
    Niizumi, Daisuke
    Takeuchi, Daiki
    Ohishi, Yasunori
    Harada, Noboru
    Kashino, Kunio
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 2391 - 2406