Reranking and Self-Training for Parser Adaptation

被引:0
|
作者
McClosky, David [1 ]
Charniak, Eugene [1 ]
Johnson, Mark [1 ]
机构
[1] Brown Univ, BLLIP, Providence, RI 02912 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Statistical parsers trained and tested on the Penn Wall Street Journal (WSJ) treebank have shown vast improvements over the last 10 years. Much of this improvement, however, is based upon an ever-increasing number of features to be trained on (typically) the WSJ treebank data. This has led to concern that such parsers may be too finely tuned to this corpus at the expense of portability to other genres. Such worries have merit. The standard "Charniak parser" checks in at a labeled precision-recall f-measure of 89.7% on the Penn WSJ test set, but only 82.9% on the test set from the Brown treebank corpus. This paper should allay these fears. In particular, we show that the reranking parser described in Charniak and Johnson (2005) improves performance of the parser on Brown to 85.2%. Furthermore, use of the self-training techniques described in (MeClosky et al., 2006) raise this to 87.8% (an error reduction of 28%) again without any use of labeled Brown data. This is remarkable since training the parser and reranker on labeled Brown data achieves only 88.4%.
引用
收藏
页码:337 / 344
页数:8
相关论文
共 50 条
  • [1] Cycle Self-Training for Domain Adaptation
    Liu, Hong
    Wang, Jianmin
    Long, Mingsheng
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [2] Self-training a Constituency Parser using N-gram Trees
    Celebi, Arda
    Ozgur, Arzucan
    [J]. LREC 2014 - NINTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2014, : 2893 - 2896
  • [3] Understanding Self-Training for Gradual Domain Adaptation
    Kumar, Ananya
    Ma, Tengyu
    Liang, Percy
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [4] Fast and Easy Sensor Adaptation With Self-Training
    Choi, Jinhyuk
    Lee, Byeongju
    Shin, Seho
    Ji, Daehyun
    [J]. IEEE ACCESS, 2023, 11 : 8870 - 8877
  • [5] Online Continual Adaptation with Active Self-Training
    Zhou, Shiji
    Zhao, Han
    Zhang, Shanghang
    Wang, Lianzhe
    Chang, Heng
    Wang, Zhi
    Zhu, Wenwu
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [6] Unsupervised Arabic Dialect Adaptation with Self-Training
    Novotney, Scott
    Schwartz, Rich
    Khudanpur, Sanjeev
    [J]. 12TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2011 (INTERSPEECH 2011), VOLS 1-5, 2011, : 548 - +
  • [7] Unsupervised domain adaptation with self-training for weed segmentation
    Huang, Yingchao
    Hussein, Amina E.
    Wang, Xin
    Bais, Abdul
    Yao, Shanshan
    Wilder, Tanis
    [J]. Intelligent Systems with Applications, 2025, 25
  • [8] Self-Training with Contrastive Learning for Adversarial Domain Adaptation
    [J]. Zhang, Xingyi (xyzhanghust@gmail.com), 1600, Institute of Electrical and Electronics Engineers Inc.
  • [9] Adversarial Domain Adaptation Enhanced via Self-training
    Altinel, Fazil
    Akkaya, Ibrahim Batuhan
    [J]. 29TH IEEE CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS (SIU 2021), 2021,
  • [10] Improve conditional adversarial domain adaptation using self-training
    Wang, Zi
    Sun, Xiaoliang
    Su, Ang
    Wang, Gang
    Li, Yang
    Yu, Qifeng
    [J]. IET IMAGE PROCESSING, 2021, 15 (10) : 2169 - 2178