Towards more effective encoders in pre-training for sequential recommendation

被引:1
|
作者
Sun, Ke [1 ]
Qian, Tieyun [1 ]
Zhong, Ming [1 ]
Li, Xuhui [2 ]
机构
[1] Wuhan Univ, Sch Comp Sci, Wuhan, Peoples R China
[2] Wuhan Univ, Sch Informat Management, Wuhan, Peoples R China
基金
中国国家自然科学基金;
关键词
Sequential recommendation; Self-supervised learning; Pre-training; Encoder; CONTEXT;
D O I
10.1007/s11280-023-01163-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Pre-training emerges as a new learning paradigm in natural language processing and computer vision. It has also been introduced into sequential recommendation in several seminal studies for alleviating data sparsity issue. However, existing methods adopt the bidirectional transformer as the encoder which suffers from two drawbacks. One is insufficient intention modeling since the transformer architecture is suitable for extracting distributed consumption intention but cannot well catch users' concentrated and occasion consumption intentions. The other is information leakage caused by foreseeing the future item in advance during the bidirectional encoding process. To address these problems, we propose to construct more effective encoders in pre-training for sequential recommendation. Specifically, we first decouple the original bidirectional process in transformer structure into two unidirectional processes which can avoid the information leakage problem and capture the distributed consumption intention. We then employ the locality-aware convolutional neural networks (CNNs) with narrow receptive field for concentrated consumption modeling. We also introduce a random shuffle strategy to empower CNN with the ability of modeling the occasion consumption. Experiments on five datasets demonstrate that our method improves the performance of various types of downstream sequential recommendation models to a large extent, and it also generates the overall better performance than the state-of-the-art self-supervised pre-training methods.
引用
收藏
页码:2801 / 2832
页数:32
相关论文
共 50 条
  • [41] Towards Masked Autoencoding Pre-training for Wide Area Motion Imagery
    Goley, Steve
    Pradhan, Rohan
    Welch, Austin
    [J]. GEOSPATIAL INFORMATICS XIII, 2023, 12525
  • [42] Towards Multi-Interest Pre-training with Sparse Capsule Network
    Tang, Zuoli
    Wang, Lin
    Zou, Lixin
    Zhang, Xiaolu
    Zhou, Jun
    Li, Chenliang
    [J]. PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 311 - 320
  • [43] Algorithms For Cold-Start Game Recommendation Based On GNN Pre-training Model
    Yang, Hongjuan
    Tian, Gang
    Xu, Chengrui
    Wang, Rui
    [J]. COMPUTER JOURNAL, 2024,
  • [44] Rethinking ImageNet Pre-training
    He, Kaiming
    Girshick, Ross
    Dollar, Piotr
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 4917 - 4926
  • [45] Pre-Training to Learn in Context
    Gu, Yuxian
    Dong, Li
    Wei, Furu
    Huang, Minlie
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 4849 - 4870
  • [46] Pre-training via Paraphrasing
    Lewis, Mike
    Ghazvininejad, Marjan
    Ghosh, Gargi
    Aghajanyan, Armen
    Wang, Sida
    Zettlemoyer, Luke
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [47] THE PRE-TRAINING SELECTION OF TEACHERS
    Barr, A. S.
    Douglas, Lois
    [J]. JOURNAL OF EDUCATIONAL RESEARCH, 1934, 28 (02): : 92 - 117
  • [48] Improving Fractal Pre-training
    Anderson, Connor
    Farrell, Ryan
    [J]. 2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 2412 - 2421
  • [49] Photo Pre-Training, But for Sketch
    Ke, L.
    Pang, Kaiyue
    Song, Yi-Zhe
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 2754 - 2764
  • [50] Pre-training phenotyping classifiers
    Dligach, Dmitriy
    Afshar, Majid
    Miller, Timothy
    [J]. JOURNAL OF BIOMEDICAL INFORMATICS, 2021, 113 (113)