Multi-Faceted Knowledge-Driven Pre-Training for Product Representation Learning

被引:2
|
作者
Zhang, Denghui [1 ]
Liu, Yanchi [4 ]
Yuan, Zixuan [2 ]
Fu, Yanjie [5 ]
Chen, Haifeng
Xiong, Hui [3 ]
机构
[1] Rutgers State Univ, Informat Syst Dept, Newark, NJ 07103 USA
[2] Rutgers State Univ, Management Sci & Informat Syst Dept, Newark, NJ 07103 USA
[3] Rutgers State Univ, Newark, NJ 07103 USA
[4] NEC Labs Amer, Princeton, NJ 08540 USA
[5] Univ Cent Florida, Dept Comp Sci, Orlando, FL 32816 USA
基金
美国国家科学基金会;
关键词
Task analysis; Monitoring; Semantics; Pediatrics; Representation learning; Electronic publishing; Electronic commerce; Product representation learning; product search; product matching; product classification; pre-trained language models;
D O I
10.1109/TKDE.2022.3200921
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As a key component of e-commerce computing, product representation learning (PRL) provides benefits for a variety of applications, including product matching, search, and categorization. The existing PRL approaches have poor language understanding ability due to their inability to capture contextualized semantics. In addition, the learned representations by existing methods are not easily transferable to new products. Inspired by the recent advance of pre-trained language models (PLMs), we make the attempt to adapt PLMs for PRL to mitigate the above issues. In this article, we develop KINDLE, a Knowledge-drIven pre-trainiNg framework for proDuct representation LEarning, which can preserve the contextual semantics and multi-faceted product knowledge robustly and flexibly. Specifically, we first extend traditional one-stage pre-training to a two-stage pre-training framework, and exploit a deliberate knowledge encoder to ensure a smooth knowledge fusion into PLM. In addition, we propose a multi-objective heterogeneous embedding method to represent thousands of knowledge elements. This helps KINDLE calibrate knowledge noise and sparsity automatically by replacing isolated classes as training targets in knowledge acquisition tasks. Furthermore, an input-aware gating network is proposed to select the most relevant knowledge for different downstream tasks. Finally, extensive experiments have demonstrated the advantages of KINDLE over the state-of-the-art baselines across three downstream tasks.
引用
收藏
页码:7239 / 7250
页数:12
相关论文
共 50 条
  • [31] GeoBERT: Pre-Training Geospatial Representation Learning on Point-of-Interest
    Gao, Yunfan
    Xiong, Yun
    Wang, Siqi
    Wang, Haofen
    APPLIED SCIENCES-BASEL, 2022, 12 (24):
  • [32] Learning Multi-faceted Activities from Heterogeneous Data with the Product Space Hierarchical Dirichlet Processes
    Thanh-Binh Nguyen
    Vu Nguyen
    Venkatesh, Svetha
    Dinh Phung
    TRENDS AND APPLICATIONS IN KNOWLEDGE DISCOVERY AND DATA MINING (PAKDD 2016), 2016, 9794 : 128 - 140
  • [33] Better Pre-Training by Reducing Representation Confusion
    Zhang, Haojie
    Liang, Mingfei
    Xie, Ruobing
    Sun, Zhenlong
    Zhang, Bo
    Lin, Leyu
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 2325 - 2336
  • [34] Discovering Representation Sprachbund For Multilingual Pre-Training
    Fan, Yimin
    Liang, Yaobo
    Muzio, Alexandre
    Hassan, Hany
    Li, Houqiang
    Zhou, Ming
    Duan, Nan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 881 - 894
  • [35] Learning to Recommend with Multi-Faceted Trust in Social Networks
    Guo, Lei
    Ma, Jun
    Chen, Zhumin
    PROCEEDINGS OF THE 22ND INTERNATIONAL CONFERENCE ON WORLD WIDE WEB (WWW'13 COMPANION), 2013, : 205 - 206
  • [36] PreQR: Pre-training Representation for SQL Understanding
    Tang, Xiu
    Wu, Sai
    Song, Mingli
    Ying, Shanshan
    Li, Feifei
    Chen, Gang
    PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA (SIGMOD '22), 2022, : 204 - 216
  • [37] A multi-faceted approach to health education and training needs in conflict
    Sabouni, A.
    Abbara, A.
    Ghadder, F.
    Qalish, D.
    Samer, J.
    EUROPEAN JOURNAL OF PUBLIC HEALTH, 2018, 28 : 61 - 61
  • [38] A multi-faceted pandemic: a review of the state of knowledge on the Zika virus
    Depoux, Anneliese
    Philibert, Aline
    Rabier, Serge
    Philippe, Henri-Jean
    Fontanet, Arnaud
    Flahault, Antoine
    PUBLIC HEALTH REVIEWS, 2018, 39
  • [39] Managing the kaleidoscope of knowledge production and sharing: a multi-faceted assignment
    Maree, Kobus
    PERSPECTIVES IN EDUCATION, 2009, 27 (03) : 207 - 210
  • [40] Multi-stage Pre-training over Simplified Multimodal Pre-training Models
    Liu, Tongtong
    Feng, Fangxiang
    Wang, Xiaojie
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 2556 - 2565