Improving Autoregressive Grammatical Error Correction with Non-autoregressive Models

被引:0
|
作者
Cao, Hang [1 ]
Cao, Zhiquan [1 ]
Hu, Chi [1 ]
Hou, Baoyu [1 ]
Xiao, Tong [1 ,2 ]
Zhu, Jingbo [1 ,2 ]
机构
[1] Northeastern Univ, NLP Lab, Sch Comp Sci & Engn, Shenyang, Peoples R China
[2] NiuTrans Res, Shenyang, Peoples R China
基金
美国国家科学基金会; 国家重点研发计划;
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Grammatical Error Correction (GEC) aims to correct grammatical errors in sentences. We find that autoregressive models tend to assign low probabilities to tokens that need corrections. Here we introduce additional signals to the training of GEC models so that these systems can learn to better predict at ambiguous positions. To do this, we use a non-autoregressive model as an auxiliary model, and develop a new regularization term of training by considering the difference in predictions between the autoregressive and non-autoregressive models. We experiment with this method on both English and Chinese GEC tasks. Experimental results show that our GEC system outperforms the baselines on all the data sets significantly.
引用
收藏
页码:12014 / 12027
页数:14
相关论文
共 50 条
  • [1] Tail-to-Tail Non-Autoregressive Sequence Prediction for Chinese Grammatical Error Correction
    Li, Piji
    Shi, Shuming
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4973 - 4984
  • [2] GEC-DePenD: Non-Autoregressive Grammatical Error Correction with Decoupled Permutation and Decoding
    Yakovlev, Konstantin
    Podolskiy, Alexander
    Bout, Andrey
    Nikolenko, Sergey
    Piontkovskaya, Irina
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 1546 - 1558
  • [3] Improving Autoregressive NMT with Non-Autoregressive Model
    Zhou, Long
    Zhang, Jiajun
    Zong, Chengqing
    WORKSHOP ON AUTOMATIC SIMULTANEOUS TRANSLATION CHALLENGES, RECENT ADVANCES, AND FUTURE DIRECTIONS, 2020, : 24 - 29
  • [4] Non-Autoregressive Speech Recognition with Error Correction Module
    Qian, Yukun
    Zhuang, Xuyi
    Zhang, Zehua
    Zhou, Lianyu
    Lin, Xu
    Wang, Mingjiang
    PROCEEDINGS OF 2022 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2022, : 1103 - 1108
  • [5] Non-Autoregressive Chinese ASR Error Correction with Phonological Training
    Fang, Zheng
    Zhang, Ruiqing
    He, Zhongjun
    Wu, Hua
    Cao, Yanan
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 5907 - 5917
  • [6] Hybrid Autoregressive and Non-Autoregressive Transformer Models for Speech Recognition
    Tian, Zhengkun
    Yi, Jiangyan
    Tao, Jianhua
    Zhang, Shuai
    Wen, Zhengqi
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 762 - 766
  • [7] Improving Non-autoregressive Machine Translation with Error Exposure and Consistency Regularization
    Chen, Xinran
    Duan, Sufeng
    Liu, Gongshen
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT III, NLPCC 2024, 2025, 15361 : 240 - 252
  • [8] Bootstrap prediction intervals for autoregressive models fitted to non-autoregressive processes
    Matteo Grigoletto
    Journal of the Italian Statistical Society, 1998, 7 (3)
  • [9] PATCorrect: Non-autoregressive Phoneme-augmented Transformer for ASR Error Correction
    Zhang, Ziji
    Wang, Zhehui
    Kamma, Rajesh
    Eswaran, Sharanya
    Sadagopan, Narayanan
    INTERSPEECH 2023, 2023, : 3904 - 3908
  • [10] Diffusion Models for Non-autoregressive Text Generation: A Survey
    Li, Yifan
    Zhou, Kun
    Zhao, Wayne Xin
    Wen, Ji-Rong
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 6692 - 6701