Multi-Faceted Knowledge-Driven Pre-Training for Product Representation Learning

被引:2
|
作者
Zhang, Denghui [1 ]
Liu, Yanchi [4 ]
Yuan, Zixuan [2 ]
Fu, Yanjie [5 ]
Chen, Haifeng
Xiong, Hui [3 ]
机构
[1] Rutgers State Univ, Informat Syst Dept, Newark, NJ 07103 USA
[2] Rutgers State Univ, Management Sci & Informat Syst Dept, Newark, NJ 07103 USA
[3] Rutgers State Univ, Newark, NJ 07103 USA
[4] NEC Labs Amer, Princeton, NJ 08540 USA
[5] Univ Cent Florida, Dept Comp Sci, Orlando, FL 32816 USA
基金
美国国家科学基金会;
关键词
Task analysis; Monitoring; Semantics; Pediatrics; Representation learning; Electronic publishing; Electronic commerce; Product representation learning; product search; product matching; product classification; pre-trained language models;
D O I
10.1109/TKDE.2022.3200921
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As a key component of e-commerce computing, product representation learning (PRL) provides benefits for a variety of applications, including product matching, search, and categorization. The existing PRL approaches have poor language understanding ability due to their inability to capture contextualized semantics. In addition, the learned representations by existing methods are not easily transferable to new products. Inspired by the recent advance of pre-trained language models (PLMs), we make the attempt to adapt PLMs for PRL to mitigate the above issues. In this article, we develop KINDLE, a Knowledge-drIven pre-trainiNg framework for proDuct representation LEarning, which can preserve the contextual semantics and multi-faceted product knowledge robustly and flexibly. Specifically, we first extend traditional one-stage pre-training to a two-stage pre-training framework, and exploit a deliberate knowledge encoder to ensure a smooth knowledge fusion into PLM. In addition, we propose a multi-objective heterogeneous embedding method to represent thousands of knowledge elements. This helps KINDLE calibrate knowledge noise and sparsity automatically by replacing isolated classes as training targets in knowledge acquisition tasks. Furthermore, an input-aware gating network is proposed to select the most relevant knowledge for different downstream tasks. Finally, extensive experiments have demonstrated the advantages of KINDLE over the state-of-the-art baselines across three downstream tasks.
引用
收藏
页码:7239 / 7250
页数:12
相关论文
共 50 条
  • [1] Multi-Faceted Knowledge-Driven Graph Neural Network for Iris Segmentation
    Wei, Jianze
    Wang, Yunlong
    Gao, Xingyu
    He, Ran
    Sun, Zhenan
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 6015 - 6027
  • [2] Boosting Video Representation Learning with Multi-Faceted Integration
    Qiu, Zhaofan
    Yao, Ting
    Ngo, Chong-Wah
    Zhang, Xiao-Ping
    Wu, Dong
    Mei, Tao
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 14025 - 14034
  • [3] Improving Knowledge Graph Representation Learning by Structure Contextual Pre-training
    Ye, Ganqiang
    Zhang, Wen
    Bi, Zhen
    Wong, Chi Man
    Chen, Hui
    Chen, Huajun
    PROCEEDINGS OF THE 10TH INTERNATIONAL JOINT CONFERENCE ON KNOWLEDGE GRAPHS (IJCKG 2021), 2021, : 151 - 155
  • [4] Multi-Faceted Route Representation Learning for Travel Time Estimation
    Liao, Tianxi
    Han, Liangzhe
    Xu, Yi
    Zhu, Tongyu
    Sun, Leilei
    Du, Bowen
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (09) : 1 - 12
  • [5] A knowledge-guided pre-training framework for improving molecular representation learning
    Li, Han
    Zhang, Ruotian
    Min, Yaosen
    Ma, Dacheng
    Zhao, Dan
    Zeng, Jianyang
    NATURE COMMUNICATIONS, 2023, 14 (01)
  • [6] Joint Pre-training and Local Re-training: Transferable Representation Learning on Multi-source Knowledge Graphs
    Sun, Zequn
    Huang, Jiacheng
    Lin, Jinghao
    Xu, Xiaozhou
    Chen, Qijin
    Hu, Wei
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 2132 - 2144
  • [7] A knowledge-guided pre-training framework for improving molecular representation learning
    Han Li
    Ruotian Zhang
    Yaosen Min
    Dacheng Ma
    Dan Zhao
    Jianyang Zeng
    Nature Communications, 14 (1)
  • [8] Multi-Faceted Rating of Product Reviews
    Baccianella, Stefano
    Esuli, Andrea
    Sebastiani, Fabrizio
    ERCIM NEWS, 2009, (77): : 60 - 61
  • [9] KMF: Knowledge-Aware Multi-Faceted Representation Learning for Zero-Shot Node Classification
    Wu, Likang
    Jiang, Junji
    Zhao, Hongke
    Wang, Hao
    Lian, Defu
    Zhang, Mengdi
    Chen, Enhong
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 2361 - 2369
  • [10] Pre-training Strategies and Datasets for Facial Representation Learning
    Bulat, Adrian
    Cheng, Shiyang
    Yang, Jing
    Garbett, Andrew
    Sanchez, Enrique
    Tzimiropoulos, Georgios
    COMPUTER VISION, ECCV 2022, PT XIII, 2022, 13673 : 107 - 125