Pre-training Mention Representations in Coreference Models

被引:0
|
作者
Varkel, Yuval [1 ]
Globerson, Amir [1 ,2 ]
机构
[1] Tel Aviv Univ, Tel Aviv, Israel
[2] Google Res, Tel Aviv, Israel
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Collecting labeled data for coreference resolution is a challenging task, requiring skilled annotators. It is thus desirable to develop coreference resolution models that can make use of unlabeled data. Here we provide such an approach for the powerful class of neural coreference models. These models rely on representations of mentions, and we show these representations can be learned in a self-supervised manner towards improving resolution accuracy. We propose two self-supervised tasks that are closely related to coreference resolution and thus improve mention representation. Applying this approach to the GAP dataset results in new state of the arts results.
引用
收藏
页码:8534 / 8540
页数:7
相关论文
共 50 条
  • [1] Advances in Pre-Training Distributed Word Representations
    Mikolov, Tomas
    Grave, Edouard
    Bojanowski, Piotr
    Puhrsch, Christian
    Joulin, Armand
    [J]. PROCEEDINGS OF THE ELEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2018), 2018, : 52 - 55
  • [2] From Uniform Models To Generic Representations: Stock Return Prediction With Pre-training
    You, Jiawei
    Han, Tianyuan
    Shen, Liping
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [3] Multi-stage Pre-training over Simplified Multimodal Pre-training Models
    Liu, Tongtong
    Feng, Fangxiang
    Wang, Xiaojie
    [J]. 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 2556 - 2565
  • [4] Realistic Channel Models Pre-training
    Huangfu, Yourui
    Wang, Jian
    Xu, Chen
    Li, Rong
    Ge, Yiqun
    Wang, Xianbin
    Zhang, Huazi
    Wang, Jun
    [J]. 2019 IEEE GLOBECOM WORKSHOPS (GC WKSHPS), 2019,
  • [5] Pre-Training Audio Representations With Self-Supervision
    Tagliasacchi, Marco
    Gfeller, Beat
    Quitry, Felix de Chaumont
    Roblek, Dominik
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2020, 27 : 600 - 604
  • [6] Better Representations via Adversarial Training in Pre-Training: A Theoretical Perspective
    Xing, Yue
    Lin, Xiaofeng
    Song, Qifan
    Xu, Yi
    Zeng, Belinda
    Cheng, Guang
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [7] GBERT: Pre-training User representations for Ephemeral Group Recommendation
    Zhang, Song
    Zheng, Nan
    Wang, Danli
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 2631 - 2639
  • [8] Contrastive Representations Pre-Training for Enhanced Discharge Summary BERT
    Won, DaeYeon
    Lee, YoungJun
    Choi, Ho-Jin
    Jung, YuChae
    [J]. 2021 IEEE 9TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI 2021), 2021, : 507 - 508
  • [9] A Method of Relation Extraction Using Pre-training Models
    Wang, Yu
    Sun, Yining
    Ma, Zuchang
    Gao, Lisheng
    Xu, Yang
    Wu, Yichen
    [J]. 2020 13TH INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DESIGN (ISCID 2020), 2020, : 176 - 179
  • [10] Improving the Sample Efficiency of Pre-training Language Models
    Berend, Gabor
    [J]. ERCIM NEWS, 2024, (136): : 38 - 40