Awakening Latent Grounding from Pretrained Language Models for Semantic Parsing

被引:0
|
作者
Liu, Qian [1 ]
Yang, Dejian [2 ]
Zhang, Jiahui [1 ]
Guo, Jiaqi [3 ]
Zhou, Bin [1 ]
Lou, Jian-Guang [2 ]
机构
[1] Beihang Univ, Beijing, Peoples R China
[2] Microsoft Res, Beijing, Peoples R China
[3] Xi An Jiao Tong Univ, Xian, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent years pretrained language models (PLMs) hit a success on several downstream tasks, showing their power on modeling language. To better understand and leverage what PLMs have learned, several techniques have emerged to explore syntactic structures entailed by PLMs. However, few efforts have been made to explore grounding capabilities of PLMs, which are also essential. In this paper, we highlight the ability of PLMs to discover which token should be grounded to which concept, if combined with our proposed erasing-then-awakening approach. Empirical studies on four datasets demonstrate that our approach can awaken latent grounding which is understandable to human experts, even if it is not exposed to such labels during training. More importantly, our approach shows great potential to benefit downstream semantic parsing models. Taking text-to-SQL as a case study, we successfully couple our approach with two off-the-shelf parsers, obtaining an absolute improvement of up to 9.8%.
引用
收藏
页码:1174 / 1189
页数:16
相关论文
共 50 条
  • [31] Controlling the Focus of Pretrained Language Generation Models
    Ji, Jiabao
    Kim, Yoon
    Glass, James
    He, Tianxing
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3291 - 3306
  • [32] Language Recognition Based on Unsupervised Pretrained Models
    Yu, Haibin
    Zhao, Jing
    Yang, Song
    Wu, Zhongqin
    Nie, Yuting
    Zhang, Wei-Qiang
    INTERSPEECH 2021, 2021, : 3271 - 3275
  • [33] Fooling MOSS Detection with Pretrained Language Models
    Biderman, Stella
    Raff, Edward
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 2933 - 2943
  • [34] Factual Consistency of Multilingual Pretrained Language Models
    Fierro, Constanza
    Sogaard, Anders
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3046 - 3052
  • [35] Pretrained Models and Evaluation Data for the Khmer Language
    Shengyi Jiang
    Sihui Fu
    Nankai Lin
    Yingwen Fu
    Tsinghua Science and Technology, 2022, 27 (04) : 709 - 718
  • [36] Pretrained Language Models for Text Generation: A Survey
    Li, Junyi
    Tang, Tianyi
    Zhao, Wayne Xin
    Wen, Ji-Rong
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4492 - 4499
  • [37] Pretrained Language Models for Sequential Sentence Classification
    Cohan, Arman
    Beltagy, Iz
    King, Daniel
    Dalvi, Bhavana
    Weld, Daniel S.
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 3693 - 3699
  • [38] Probing Pretrained Language Models with Hierarchy Properties
    Lovon-Melgarejo, Jesus
    Moreno, Jose G.
    Besancon, Romaric
    Ferret, Olivier
    Tamine, Lynda
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT II, 2024, 14609 : 126 - 142
  • [39] Eliciting Knowledge from Pretrained Language Models for Prototypical Prompt Verbalizer
    Wei, Yinyi
    Mo, Tong
    Jiang, Yongtao
    Li, Weiping
    Zhao, Wen
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT II, 2022, 13530 : 222 - 233
  • [40] Probing Pretrained Language Models for Lexical Semantics
    Vulie, Ivan
    Ponti, Edoardo M.
    Litschko, Robert
    Glava, Goran
    Korhonen, Anna
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 7222 - 7240