Pre-training Mention Representations in Coreference Models

被引:0
|
作者
Varkel, Yuval [1 ]
Globerson, Amir [1 ,2 ]
机构
[1] Tel Aviv Univ, Tel Aviv, Israel
[2] Google Res, Tel Aviv, Israel
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Collecting labeled data for coreference resolution is a challenging task, requiring skilled annotators. It is thus desirable to develop coreference resolution models that can make use of unlabeled data. Here we provide such an approach for the powerful class of neural coreference models. These models rely on representations of mentions, and we show these representations can be learned in a self-supervised manner towards improving resolution accuracy. We propose two self-supervised tasks that are closely related to coreference resolution and thus improve mention representation. Applying this approach to the GAP dataset results in new state of the arts results.
引用
收藏
页码:8534 / 8540
页数:7
相关论文
共 50 条
  • [21] Revisiting Weakly Supervised Pre-Training of Visual Perception Models
    Singh, Mannat
    Gustafson, Laura
    Adcock, Aaron
    Reis, Vinicius De Freitas
    Gedik, Bugra
    Kosaraju, Raj Prateek
    Mahajan, Dhruv
    Girshick, Ross
    Dollar, Piotr
    Van Der Maaten, Laurens
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 794 - 804
  • [22] Pre-training Summarization Models of Structured Datasets for Cardinality Estimation
    Lu, Yao
    Kandula, Srikanth
    Konig, Arnd Christian
    Chaudhuri, Surajit
    [J]. PROCEEDINGS OF THE VLDB ENDOWMENT, 2021, 15 (03): : 414 - 426
  • [23] Progress in protein pre-training models integrating structural knowledge
    Tang, Tian-Yi
    Xiong, Yi-Ming
    Zhang, Rui-Ge
    Zhang, Jian
    Li, Wen-Fei
    Wang, Jun
    Wang, Wei
    [J]. Wuli Xuebao/Acta Physica Sinica, 2024, 73 (18):
  • [24] Improving Image Representations via MoCo Pre-training for Multimodal CXR Classification
    Serra, Francesco Dalla
    Jacenkow, Grzegorz
    Deligianni, Fani
    Dalton, Jeff
    O'Neil, Alison Q.
    [J]. MEDICAL IMAGE UNDERSTANDING AND ANALYSIS, MIUA 2022, 2022, 13413 : 623 - 635
  • [25] Pre-Training Transformers as Energy-Based Cloze Models
    Clark, Kevin
    Luong, Minh-Thang
    Le, Quoc V.
    Manning, Christopher D.
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 285 - 294
  • [26] Research on Pre-Training Models for Tibetan Text with Character Awareness
    Gadeng, Luosang
    Nyima, Tashi
    [J]. Computer Engineering and Applications, 2024, 60 (21) : 127 - 133
  • [27] Removing Backdoors in Pre-trained Models by Regularized Continual Pre-training
    Zhu, Biru
    Cui, Ganqu
    Chen, Yangyi
    Qin, Yujia
    Yuan, Lifan
    Fu, Chong
    Deng, Yangdong
    Liu, Zhiyuan
    Sun, Maosong
    Gu, Ming
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2023, 11 : 1608 - 1623
  • [28] Supervised Contrastive Pre-training for Mammographic Triage Screening Models
    Cao, Zhenjie
    Yang, Zhicheng
    Tang, Yuxing
    Zhang, Yanbo
    Han, Mei
    Xiao, Jing
    Ma, Jie
    Chang, Peng
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT VII, 2021, 12907 : 129 - 139
  • [29] UER: An Open-Source Toolkit for Pre-training Models
    Zhao, Zhe
    Chen, Hui
    Zhang, Jinbin
    Zhao, Xin
    Liu, Tao
    Lu, Wei
    Chen, Xi
    Deng, Haotang
    Ju, Qi
    Du, Xiaoyong
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF SYSTEM DEMONSTRATIONS, 2019, : 241 - 246
  • [30] On the importance of pre-training data volume for compact language models
    Micheli, Vincent
    D'Hoffschmidt, Martin
    Fleuret, Francois
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 7853 - 7858