Towards more effective encoders in pre-training for sequential recommendation

被引:1
|
作者
Sun, Ke [1 ]
Qian, Tieyun [1 ]
Zhong, Ming [1 ]
Li, Xuhui [2 ]
机构
[1] Wuhan Univ, Sch Comp Sci, Wuhan, Peoples R China
[2] Wuhan Univ, Sch Informat Management, Wuhan, Peoples R China
来源
WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS | 2023年 / 26卷 / 05期
基金
中国国家自然科学基金;
关键词
Sequential recommendation; Self-supervised learning; Pre-training; Encoder; CONTEXT;
D O I
10.1007/s11280-023-01163-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Pre-training emerges as a new learning paradigm in natural language processing and computer vision. It has also been introduced into sequential recommendation in several seminal studies for alleviating data sparsity issue. However, existing methods adopt the bidirectional transformer as the encoder which suffers from two drawbacks. One is insufficient intention modeling since the transformer architecture is suitable for extracting distributed consumption intention but cannot well catch users' concentrated and occasion consumption intentions. The other is information leakage caused by foreseeing the future item in advance during the bidirectional encoding process. To address these problems, we propose to construct more effective encoders in pre-training for sequential recommendation. Specifically, we first decouple the original bidirectional process in transformer structure into two unidirectional processes which can avoid the information leakage problem and capture the distributed consumption intention. We then employ the locality-aware convolutional neural networks (CNNs) with narrow receptive field for concentrated consumption modeling. We also introduce a random shuffle strategy to empower CNN with the ability of modeling the occasion consumption. Experiments on five datasets demonstrate that our method improves the performance of various types of downstream sequential recommendation models to a large extent, and it also generates the overall better performance than the state-of-the-art self-supervised pre-training methods.
引用
收藏
页码:2801 / 2832
页数:32
相关论文
共 50 条
  • [21] RESETBERT4Rec: A Pre-training Model Integrating Time And User Historical Behavior for Sequential Recommendation
    Zhao, Qihang
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 1812 - 1816
  • [22] Emerging Property of Masked Token for Effective Pre-training
    Choi, Hyesong
    Lee, Hunsang
    Joung, Seyoung
    Park, Hyejin
    Kim, Jiyeong
    Min, Dongbo
    COMPUTER VISION - ECCV 2024, PT LXXVI, 2025, 15134 : 272 - 289
  • [23] SEPT: Towards Scalable and Efficient Visual Pre-training
    Lin, Yiqi
    Zheng, Huabin
    Zhong, Huaping
    Zhu, Jinjing
    Li, Weijia
    He, Conghui
    Wang, Lin
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 1622 - 1630
  • [24] Task2Sim: Towards Effective Pre-training and Transfer from Synthetic Data
    Mishra, Samarth
    Panda, Rameswar
    Phoo, Cheng Perng
    Chen, Chun-Fu
    Karlinsky, Leonid
    Saenko, Kate
    Saligrama, Venkatesh
    Feris, Rogerio S.
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9184 - 9194
  • [25] U-BERT: Pre-training User Representations for Improved Recommendation
    Qiu, Zhaopeng
    Wu, Xian
    Gao, Jingyue
    Fan, Wei
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 4320 - 4327
  • [26] Medication Recommendation Based on a Knowledge-enhanced Pre-training Model
    Wang, Mengzhen
    Chen, Jianhui
    Lin, Shaofu
    PROCEEDINGS OF 2021 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY WORKSHOPS AND SPECIAL SESSIONS: (WI-IAT WORKSHOP/SPECIAL SESSION 2021), 2021, : 290 - 294
  • [27] Towards a Holistic Understanding of Mathematical Questions with Contrastive Pre-training
    Ning, Yuting
    Huang, Zhenya
    Lin, Xin
    Chen, Enhong
    Tong, Shiwei
    Gong, Zheng
    Wang, Shijin
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 13409 - 13418
  • [28] EnCore: Fine-Grained Entity Typing by Pre-Training Entity Encoders on Coreference Chains
    Mtumbuka, Frank
    Schockaert, Steven
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 1768 - 1781
  • [29] Bipartite Graph Pre-training for Unsupervised Extractive Summarization with Graph Convolutional Auto-Encoders
    Mao, Qianren
    Zhao, Shaobo
    Li, Jiarui
    Gu, Xiaolei
    He, Shizhu
    Li, Bo
    Li, Jianxin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 4929 - 4941
  • [30] Heterogeneous graph convolutional network pre-training as side information for improving recommendation
    Do, Phuc
    Pham, Phu
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (18): : 15945 - 15961