KEEP: An Industrial Pre-Training Framework for Online Recommendation via Knowledge Extraction and Plugging

被引:9
|
作者
Zhang, Yujing [1 ]
Chan, Zhangming [1 ]
Xu, Shuhao [2 ]
Bian, Weijie [1 ]
Han, Shuguang [1 ]
Deng, Hongbo [1 ]
Zheng, Bo [1 ]
机构
[1] Alibaba Grp, Beijing, Peoples R China
[2] Tsinghua Univ, Sch Software, Beijing, Peoples R China
关键词
Online Recommendation; Pre-training; Knowledge Extraction; Knowledge Plugging;
D O I
10.1145/3511808.3557106
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
An industrial recommender system generally presents a hybrid list that contains results from multiple subsystems. In practice, each subsystem is optimized with its own feedback data to avoid the disturbance among different subsystems. However, we argue that such data usage may lead to sub-optimal online performance because of the data sparsity. To alleviate this issue, we propose to extract knowledge from the super-domain that contains webscale and long-time impression data, and further assist the online recommendation task (downstream task). To this end, we propose a novel industrial KnowlEdge Extraction and Plugging (KEEP) framework, which is a two-stage framework that consists of 1) a supervised pre-training knowledge extraction module on superdomain, and 2) a plug-in network that incorporates the extracted knowledge into the downstream model. This makes it friendly for incremental training of online recommendation. Moreover, we design an efficient empirical approach for KEEP and introduce our hands-on experience during the implementation of KEEP in a largescale industrial system. Experiments conducted on two real-world datasets demonstrate that KEEP can achieve promising results. It is notable that KEEP has also been deployed on the display advertising system in Alibaba, bringing a lift of +5.4% CTR and +4.7% RPM.
引用
收藏
页码:3684 / 3693
页数:10
相关论文
共 50 条
  • [1] Knowledge Transfer via Pre-training for Recommendation: A Review and Prospect
    Zeng, Zheni
    Xiao, Chaojun
    Yao, Yuan
    Xie, Ruobing
    Liu, Zhiyuan
    Lin, Fen
    Lin, Leyu
    Sun, Maosong
    [J]. FRONTIERS IN BIG DATA, 2021, 4
  • [2] Medication Recommendation Based on a Knowledge-enhanced Pre-training Model
    Wang, Mengzhen
    Chen, Jianhui
    Lin, Shaofu
    [J]. PROCEEDINGS OF 2021 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY WORKSHOPS AND SPECIAL SESSIONS: (WI-IAT WORKSHOP/SPECIAL SESSION 2021), 2021, : 290 - 294
  • [3] Improving Knowledge Tracing via Pre-training Question Embeddings
    Liu, Yunfei
    Yang, Yang
    Chen, Xianyu
    Shen, Jian
    Zhang, Haifeng
    Yu, Yong
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 1577 - 1583
  • [4] Improving Knowledge Tracing via Pre-training Question Embeddings
    Liu, Yunfei
    Yang, Yang
    Chen, Xianyu
    Shen, Jian
    Zhang, Haifeng
    Yu, Yong
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 1556 - 1562
  • [5] Temporal Contrastive Pre-Training for Sequential Recommendation
    Tian, Changxin
    Lin, Zihan
    Bian, Shuqing
    Wang, Jinpeng
    Zhao, Wayne Xin
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 1925 - 1934
  • [6] Improving News Recommendation via Bottlenecked Multi-task Pre-training
    Xiao, Xiongfeng
    Li, Qing
    Liu, Songlin
    Zhou, Kun
    [J]. PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2082 - 2086
  • [7] Pre-training via Paraphrasing
    Lewis, Mike
    Ghazvininejad, Marjan
    Ghosh, Gargi
    Aghajanyan, Armen
    Wang, Sida
    Zettlemoyer, Luke
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [8] Friend Ranking in Online Games via Pre-training Edge Transformers
    Yao, Liang
    Peng, Jiazhen
    Ji, Shenggong
    Liu, Qiang
    Cai, Hongyun
    He, Feng
    [J]. PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2016 - 2020
  • [9] Lightweight Model Pre-Training via Language Guided Knowledge Distillation
    Li, Mingsheng
    Zhang, Lin
    Zhu, Mingzhen
    Huang, Zilong
    Yu, Gang
    Fan, Jiayuan
    Chen, Tao
    [J]. IEEE Transactions on Multimedia, 2024, 26 : 10720 - 10730
  • [10] Multi-Modal Contrastive Pre-training for Recommendation
    Liu, Zhuang
    Ma, Yunpu
    Schubert, Matthias
    Ouyang, Yuanxin
    Xiong, Zhang
    [J]. PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2022, 2022, : 99 - 108