Evaluating Pre-training Strategies for Collaborative Filtering

被引:0
|
作者
da Costa, Julio B. G. [1 ]
Marinho, Leandro B. [1 ]
Santos, Rodrygo L. T. [2 ]
Parra, Denis [3 ]
机构
[1] Univ Fed Campina Grande, Campina Grande, Paraiba, Brazil
[2] Univ Fed Minas Gerais, Belo Horizonte, MG, Brazil
[3] PUC Chile, Santiago, Chile
来源
2023 PROCEEDINGS OF THE 31ST ACM CONFERENCE ON USER MODELING, ADAPTATION AND PERSONALIZATION, UMAP 2023 | 2023年
关键词
model initialization; transfer learning; collaborative filtering;
D O I
10.1145/3565472.3592949
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Pre-training is essential for effective representation learning models, especially in natural language processing and computer vision-related tasks. The core idea is to learn representations, usually through unsupervised or self-supervised approaches on large and generic source datasets, and use those pre-trained representations (aka embeddings) as initial parameter values during training on the target dataset. Seminal works in this area show that pre-training can act as a regularization mechanism placing the model parameters in regions of the optimization landscape closer to better local minima than random parameter initialization. However, no systematic studies evaluate the effectiveness of pre-training strategies on model-based collaborative filtering. This paper conducts a broad set of experiments to evaluate different pre-training strategies for collaborative filtering using Matrix Factorization (MF) as the base model. We show that such models equipped with pre-training in a transfer learning setting can vastly improve the prediction quality compared to the standard random parameter initialization baseline, reaching state-of-the-art results in standard recommender systems benchmarks. We also present alternatives for the out-of-vocabulary item problem (i.e., items present in target but not in source datasets) and show that pre-training in the context of MF acts as a regularizer, explaining the improvement in model generalization.
引用
收藏
页码:175 / 182
页数:8
相关论文
共 50 条
  • [1] An Adaptive Graph Pre-training Framework for Localized Collaborative Filtering
    Wang, Yiqi
    Li, Chaozhuo
    Liu, Zheng
    Li, Mingzheng
    Tang, Jiliang
    Xie, Xing
    Chen, Lei
    Yu, Philip S.
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2023, 41 (02)
  • [2] Graph neural collaborative filtering with medical content-aware pre-training for treatment pattern recommendation
    Min, Xin
    Li, Wei
    Han, Ruiqi
    Ji, Tianlong
    Xie, Weidong
    PATTERN RECOGNITION LETTERS, 2024, 185 : 210 - 217
  • [3] Evaluating synthetic pre-Training for handwriting processing tasks
    Pippi, Vittorio
    Cascianelli, Silvia
    Baraldi, Lorenzo
    Cucchiara, Rita
    PATTERN RECOGNITION LETTERS, 2023, 172 : 44 - 50
  • [4] Pre-training Strategies and Datasets for Facial Representation Learning
    Bulat, Adrian
    Cheng, Shiyang
    Yang, Jing
    Garbett, Andrew
    Sanchez, Enrique
    Tzimiropoulos, Georgios
    COMPUTER VISION, ECCV 2022, PT XIII, 2022, 13673 : 107 - 125
  • [5] Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training
    Radenovic, Filip
    Dubey, Abhimanyu
    Kadian, Abhishek
    Mihaylov, Todor
    Vandenhende, Simon
    Patel, Yash
    Wen, Yi
    Ramanathan, Vignesh
    Mahajan, Dhruv
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 6967 - 6977
  • [6] Unsupervised Video Domain Adaptation with Masked Pre-Training and Collaborative Self-Training
    Reddy, Arun
    Paul, William
    Rivera, Corban
    Shah, Ketul
    de Melo, Celso M.
    Chellappa, Rama
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 18919 - 18929
  • [7] Evaluating the Use of Synthetic Queries for Pre-training a Semantic Query Tagger
    Bassani, Elias
    Pasi, Gabriella
    ADVANCES IN INFORMATION RETRIEVAL, PT II, 2022, 13186 : 39 - 46
  • [8] Pre-training and Evaluating Transformer-based Language Models for Icelandic
    Daoason, Jon Friorik
    Loftsson, Hrafn
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 7386 - 7391
  • [9] A Study into Pre-training Strategies for Spoken Language Understanding on Dysarthric Speech
    Wang, Pu
    BabaAli, Bagher
    Van Hamme, Hugo
    INTERSPEECH 2021, 2021, : 36 - 40
  • [10] Multi-stage Pre-training over Simplified Multimodal Pre-training Models
    Liu, Tongtong
    Feng, Fangxiang
    Wang, Xiaojie
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 2556 - 2565