QUERT: Continual Pre-training of Language Model for Query Understanding in Travel Domain Search

被引:0
|
作者
Xie, Jian [1 ]
Liang, Yidan [2 ]
Liu, Jingping [3 ]
Xiao, Yanghua [1 ]
Wu, Baohua [2 ]
Ni, Shenghua [2 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai Key Lab Data Sci, Shanghai, Peoples R China
[2] Alibaba Grp, Hangzhou, Peoples R China
[3] East China Univ Sci & Technol, Sch Informat Sci & Engn, Shanghai, Peoples R China
关键词
Continual Pre-training; Query Understanding; Travel Domain Search;
D O I
10.1145/3580305.3599891
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In light of the success of the pre-trained language models (PLMs), continual pre-training of generic PLMs has been the paradigm of domain adaption. In this paper, we propose QUERT, A Continual Pre-trained Language Model for QUERy Understanding in Travel Domain Search. QUERT is jointly trained on four tailored pre-training tasks to the characteristics of query in travel domain search: Geography-aware Mask Prediction, Geohash Code Prediction, User Click Behavior Learning, and Phrase and Token Order Prediction. Performance improvement of downstream tasks and ablation experiment demonstrate the effectiveness of our proposed pre-training tasks. To be specific, the average performance of downstream tasks increases by 2.02% and 30.93% in supervised and unsupervised settings, respectively. To check on the improvement of QUERT to online business, we deploy QUERT and perform A/B testing on Fliggy APP. The feedback results show that QUERT increases the Unique Click-Through Rate and Page Click-Through Rate by 0.89% and 1.03% when applying QUERT as the encoder. Resources are available at https://github.com/hsaest/QUERT.
引用
收藏
页码:5282 / 5291
页数:10
相关论文
共 50 条
  • [31] Pre-training and Evaluation of Numeracy-oriented Language Model
    Feng, Fuli
    Rui, Xilin
    Wang, Wenjie
    Cao, Yixin
    Chua, Tat-Seng
    ICAIF 2021: THE SECOND ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, 2021,
  • [32] Pre-training A Prompt Pool for Vision-Language Model
    Liu, Jun
    Gu, Yang
    Yang, Zhaohua
    Guo, Shuai
    Liu, Huaqiu
    Chen, Yiqiang
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [33] REALM: Retrieval-Augmented Language Model Pre-Training
    Guu, Kelvin
    Lee, Kenton
    Tung, Zora
    Pasupat, Panupong
    Chang, Ming-Wei
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [34] Understanding tables with intermediate pre-training
    Eisenschlos, Julian Martin
    Krichene, Syrine
    Mueller, Thomas
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020,
  • [35] oLMpics-On What Language Model Pre-training Captures
    Talmor, Alon
    Elazar, Yanai
    Goldberg, Yoav
    Berant, Jonathan
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2020, 8 (08) : 743 - 758
  • [36] Gradual Syntactic Label Replacement for Language Model Pre-Training
    Wang, Yile
    Zhang, Yue
    Li, Peng
    Liu, Yang
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 486 - 496
  • [37] Learning Better Masking for Better Language Model Pre-training
    Yang, Dongjie
    Zhang, Zhuosheng
    Zhao, Hai
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 7255 - 7267
  • [38] Pre-training Universal Language Representation
    Li, Yian
    Zhao, Hai
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 5122 - 5133
  • [39] Multi-Grained Topological Pre-Training of Language Models in Sponsored Search
    Tian, Zhoujin
    Li, Chaozhuo
    Zuo, Zhiqiang
    Wen, Zengxuan
    Hu, Xinyue
    Han, Xiao
    Huang, Haizhen
    Wang, Senzhang
    Deng, Weiwei
    Xie, Xing
    Zhang, Qi
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2189 - 2193
  • [40] K-DLM: A Domain-Adaptive Language Model Pre-Training Framework with Knowledge Graph
    Zou, Jiaxin
    Xie, Zuotong
    Chen, Junhua
    Hou, Jiawei
    Yan, Qiang
    Zheng, Hai-Tao
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT IV, 2023, 14257 : 447 - 459