PEPT: Expert Finding Meets Personalized Pre-Training

被引:0
|
作者
Peng, Qiyao [1 ]
Xu, Hongyan [2 ]
Wang, Yinghui [3 ]
Liu, Hongtao [4 ]
Huo, Cuiying [5 ]
Wang, Wenjun [5 ,6 ]
机构
[1] Tianjin Univ, Sch New Media & Commun, Tianjin, Peoples R China
[2] Tianjin Univ, Coll Intelligence & Comp, Tianjin, Peoples R China
[3] Beijing Inst Control & Elect Technol, Key Lab Informat Syst & Technol, Beijing, Peoples R China
[4] Du Xiaoman Technol, Beijing, Peoples R China
[5] Tianjin Univ, Coll Intelligence & Comp, Tianjin, Peoples R China
[6] Hainan Trop Ocean Univ, YazhouBay Innovat Inst, Hainan, Peoples R China
关键词
Contrastive Learning;
D O I
10.1145/3690380
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Finding experts is essential in Community Question Answering (CQA) platforms as it enables the effective routing of questions to potential users who can provide relevant answers. The key is to personalized learning expert representations based on their historical answered questions, and accurately matching them with target questions. Recently, the applications of Pre-Trained Language Models (PLMs) have gained significant attraction due to their impressive capability to comprehend textual data, and are widespread used across various domains. There have been some preliminary works exploring the usability of PLMs in expert finding, such as pre-training expert or question representations. However, these models usually learn pure text representations of experts from histories, disregarding personalized and fine-grained expert modeling. For alleviating this, we present a personalized pre-training and fine-tuning paradigm, which could effectively learn expert interest and expertise simultaneously. Specifically, in our pre-training framework, we integrate historical answered questions of one expert with one target question, and regard it as a candidate-aware expert-level input unit. Then, we fuse expert IDs into the pre-training for guiding the model to model personalized expert representations, which can help capture the unique characteristics and expertise of each individual expert. Additionally, in our pre-training task, we design (1) a question-level masked language model task to learn the relatedness between histories, enabling the modeling of question-level expert interest; (2) a vote-oriented task to capture question-level expert expertise by predicting the vote score the expert would receive. Through our pre-training framework and tasks, our approach could holistically learn expert representations including interests and expertise. Our method has been extensively evaluated on six real-world CQA datasets, and the experimental results consistently demonstrate the superiority of our approach over competitive baseline methods.
引用
收藏
页数:26
相关论文
共 50 条
  • [21] Rethinking Pre-training and Self-training
    Zoph, Barret
    Ghiasi, Golnaz
    Lin, Tsung-Yi
    Cui, Yin
    Liu, Hanxiao
    Cubuk, Ekin D.
    Le, Quoc V.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [22] PerFedRec plus plus : Enhancing Personalized Federated Recommendation with Self-Supervised Pre-Training
    Luo, Sichun
    Xiao, Yuanzhang
    Zhang, Xinyi
    Liu, Yang
    Ding, Wenbo
    Song, Linqi
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (05)
  • [23] Pre-Training Without Natural Images
    Kataoka, Hirokatsu
    Okayasu, Kazushige
    Matsumoto, Asato
    Yamagata, Eisuke
    Yamada, Ryosuke
    Inoue, Nakamasa
    Nakamura, Akio
    Satoh, Yutaka
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (04) : 990 - 1007
  • [24] A Pipelined Pre-training Algorithm for DBNs
    Ma, Zhiqiang
    Li, Tuya
    Yang, Shuangtao
    Zhang, Li
    CHINESE COMPUTATIONAL LINGUISTICS AND NATURAL LANGUAGE PROCESSING BASED ON NATURALLY ANNOTATED BIG DATA, CCL 2017, 2017, 10565 : 48 - 59
  • [25] Dialogue-oriented Pre-training
    Xu, Yi
    Zhao, Hai
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 2663 - 2673
  • [26] Improving fault localization with pre-training
    Zhang, Zhuo
    Li, Ya
    Xue, Jianxin
    Mao, Xiaoguang
    FRONTIERS OF COMPUTER SCIENCE, 2024, 18 (01)
  • [27] Robot Learning with Sensorimotor Pre-training
    Radosavovic, Ilija
    Shi, Baifeng
    Fu, Letian
    Goldberg, Ken
    Darrell, Trevor
    Malik, Jitendra
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [28] Simulated SAR for ATR pre-training
    Willis, Christopher J.
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING IN DEFENSE APPLICATIONS III, 2021, 11870
  • [29] Structural Pre-training for Dialogue Comprehension
    Zhang, Zhuosheng
    Zhao, Hai
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 5134 - 5145
  • [30] Pre-Training Methods for Question Reranking
    Campese, Stefano
    Lauriola, Ivano
    Moschitti, Alessandro
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 469 - 476