MCM: A Multi-task Pre-trained Customer Model for Personalization

被引:0
|
作者
Luo, Rui [1 ]
Wang, Tianxin [1 ]
Deng, Jingyuan [1 ]
Wan, Peng [1 ]
机构
[1] Amazon LLC, Beijing, Peoples R China
关键词
personalization; transformer; multi-task; recommendation; pretrain;
D O I
10.1145/3604915.3608868
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Personalization plays a critical role in helping customers discover the products and contents they prefer for e-commerce stores.Personalized recommendations differ in contents, target customers, and UI. However, they require a common core capability - the ability to deeply understand customers' preferences and shopping intents. In this paper, we introduce the MCM (Multitask pre-trained Customer Model), a large pre-trained BERT-based multi-task customer model with 10 million trainable parameters for e-commerce stores. This model aims to empower all personalization projects by providing commonly used preference scores for recommendations, customer embeddings for transfer learning, and a pretrained model for fine-tuning. In this work, we improve the SOTA BERT4Rec framework to handle heterogeneous customer signals and multi-task training as well as innovate new data augmentation method that is suitable for recommendation task. Experimental results show that MCM outperforms the original BERT4Rec by 17% on on NDCG@10 of next action prediction tasks. Additionally, we demonstrate that the model can be easily fine-tuned to assist a specific recommendation task. For instance, after fine-tuning MCM for an incentive based recommendation project, performance improves by 60% on the conversion prediction task and 25% on the click-through prediction task compared to a baseline tree-based GBDT model.
引用
收藏
页码:637 / 639
页数:3
相关论文
共 50 条
  • [1] Multi-task Learning based Pre-trained Language Model for Code Completion
    Liu, Fang
    Li, Ge
    Zhao, Yunfei
    Jin, Zhi
    [J]. 2020 35TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING (ASE 2020), 2020, : 473 - 485
  • [2] Drug knowledge discovery via multi-task learning and pre-trained models
    Li, Dongfang
    Xiong, Ying
    Hu, Baotian
    Tang, Buzhou
    Peng, Weihua
    Chen, Qingcai
    [J]. BMC MEDICAL INFORMATICS AND DECISION MAKING, 2021, 21 (SUPPL 9)
  • [3] Multi-task Active Learning for Pre-trained Transformer-based Models
    Rotman, Guy
    Reichart, Roi
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 1209 - 1228
  • [4] Enhancing Pre-trained Language Representation for Multi-Task Learning of Scientific Summarization
    Jia, Ruipeng
    Cao, Yannan
    Fang, Fang
    Li, Jinpeng
    Liu, Yanbing
    Yin, Pengfei
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [5] Drug knowledge discovery via multi-task learning and pre-trained models
    Dongfang Li
    Ying Xiong
    Baotian Hu
    Buzhou Tang
    Weihua Peng
    Qingcai Chen
    [J]. BMC Medical Informatics and Decision Making, 21
  • [6] JiuZhang 2.0: A Unified Chinese Pre-trained Language Model for Multi-task Mathematical Problem Solving
    Zhao, Wayne Xin
    Zhou, Kun
    Zhang, Beichen
    Gong, Zheng
    Chen, Zhipeng
    Zhou, Yuanhang
    Wen, Ji-Rong
    Sha, Jing
    Wang, Shijin
    Liu, Cong
    Hu, Guoping
    [J]. PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 5660 - 5672
  • [7] MASTER: Multi-task Pre-trained Bottlenecked Masked Autoencoders Are Better Dense Retrievers
    Zhou, Kun
    Liu, Xiao
    Gong, Yeyun
    Zhao, Wayne Xin
    Jiang, Daxin
    Duan, Nan
    Wen, Ji-Rong
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT II, 2023, 14170 : 630 - 647
  • [8] Multi-task Learning Based Online Dialogic Instruction Detection with Pre-trained Language Models
    Hao, Yang
    Li, Hang
    Ding, Wenbiao
    Wu, Zhongqin
    Tang, Jiliang
    Luckin, Rose
    Liu, Zitao
    [J]. ARTIFICIAL INTELLIGENCE IN EDUCATION (AIED 2021), PT II, 2021, 12749 : 183 - 189
  • [9] MutaPT: A Multi-Task Pre-Trained Transformer for Classifying State of Disorders of Consciousness Using EEG Signal
    Wang, Zihan
    Yu, Junqi
    Gao, Jiahui
    Bai, Yang
    Wan, Zhijiang
    [J]. BRAIN SCIENCES, 2024, 14 (07)
  • [10] PTMB: An online satellite task scheduling framework based on pre-trained Markov decision process for multi-task scenario
    Li, Guohao
    Li, Xuefei
    Li, Jing
    Chen, Jia
    Shen, Xin
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 284