Reranking and Self-Training for Parser Adaptation

被引:0
|
作者
McClosky, David [1 ]
Charniak, Eugene [1 ]
Johnson, Mark [1 ]
机构
[1] Brown Univ, BLLIP, Providence, RI 02912 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Statistical parsers trained and tested on the Penn Wall Street Journal (WSJ) treebank have shown vast improvements over the last 10 years. Much of this improvement, however, is based upon an ever-increasing number of features to be trained on (typically) the WSJ treebank data. This has led to concern that such parsers may be too finely tuned to this corpus at the expense of portability to other genres. Such worries have merit. The standard "Charniak parser" checks in at a labeled precision-recall f-measure of 89.7% on the Penn WSJ test set, but only 82.9% on the test set from the Brown treebank corpus. This paper should allay these fears. In particular, we show that the reranking parser described in Charniak and Johnson (2005) improves performance of the parser on Brown to 85.2%. Furthermore, use of the self-training techniques described in (MeClosky et al., 2006) raise this to 87.8% (an error reduction of 28%) again without any use of labeled Brown data. This is remarkable since training the parser and reranker on labeled Brown data achieves only 88.4%.
引用
收藏
页码:337 / 344
页数:8
相关论文
共 50 条
  • [41] ADVERSARIAL SELF-TRAINING UNSUPERVISED DOMAIN ADAPTATION FOR REMOTE SENSING SCENE CLASSIFICATION
    Huang, Ying
    Li, Tangsheng
    Liu, Su
    Mei, Wenhao
    2022 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2022), 2022, : 1572 - 1575
  • [42] SETRED: Self-training with editing
    Li, M
    Zhou, ZH
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PROCEEDINGS, 2005, 3518 : 611 - 621
  • [43] DaMSTF: Domain Adversarial Learning Enhanced Meta Self-Training for Domain Adaptation
    Lu, Menglong
    Huang, Zhen
    Zhao, Yunxiang
    Tian, Zhiliang
    Liu, Yang
    Li, Dongsheng
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 1650 - 1668
  • [44] Deep Bayesian Self-Training
    Fabio De Sousa Ribeiro
    Francesco Calivá
    Mark Swainson
    Kjartan Gudmundsson
    Georgios Leontidis
    Stefanos Kollias
    Neural Computing and Applications, 2020, 32 : 4275 - 4291
  • [45] Confidence Regularized Self-Training
    Zou, Yang
    Yu, Zhiding
    Liu, Xiaofeng
    Kumar, B. V. K. Vijaya
    Wang, Jinsong
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 5981 - 5990
  • [46] Self-Training with Weak Supervision
    Karamanolakis, Giannis
    Mukherjee, Subhabrata
    Zheng, Guoqing
    Awadallah, Ahmed Hassan
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 845 - 863
  • [48] KUDOS FOR SELF-TRAINING AIDS
    BRYANT, SF
    COMPUTER DECISIONS, 1984, 16 (14): : 44 - &
  • [49] Doubly Robust Self-Training
    Zhu, Banghua
    Ding, Mingyu
    Jacobson, Philip
    Wu, Ming
    Zhan, Wei
    Jordan, Michael I.
    Jiao, Jiantao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [50] Deep Bayesian Self-Training
    Ribeiro, Fabio De Sousa
    Caliva, Francesco
    Swainson, Mark
    Gudmundsson, Kjartan
    Leontidis, Georgios
    Kollias, Stefanos
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (09): : 4275 - 4291