Awakening Latent Grounding from Pretrained Language Models for Semantic Parsing

被引:0
|
作者
Liu, Qian [1 ]
Yang, Dejian [2 ]
Zhang, Jiahui [1 ]
Guo, Jiaqi [3 ]
Zhou, Bin [1 ]
Lou, Jian-Guang [2 ]
机构
[1] Beihang Univ, Beijing, Peoples R China
[2] Microsoft Res, Beijing, Peoples R China
[3] Xi An Jiao Tong Univ, Xian, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent years pretrained language models (PLMs) hit a success on several downstream tasks, showing their power on modeling language. To better understand and leverage what PLMs have learned, several techniques have emerged to explore syntactic structures entailed by PLMs. However, few efforts have been made to explore grounding capabilities of PLMs, which are also essential. In this paper, we highlight the ability of PLMs to discover which token should be grounded to which concept, if combined with our proposed erasing-then-awakening approach. Empirical studies on four datasets demonstrate that our approach can awaken latent grounding which is understandable to human experts, even if it is not exposed to such labels during training. More importantly, our approach shows great potential to benefit downstream semantic parsing models. Taking text-to-SQL as a case study, we successfully couple our approach with two off-the-shelf parsers, obtaining an absolute improvement of up to 9.8%.
引用
收藏
页码:1174 / 1189
页数:16
相关论文
共 50 条
  • [41] Pretrained models and evaluation data for the Khmer language
    Jiang, Shengyi
    Fu, Sihui
    Lin, Nankai
    Fu, Yingwen
    TSINGHUA SCIENCE AND TECHNOLOGY, 2022, 27 (04) : 709 - 718
  • [42] Differentiable Parsing and Visual Grounding of Natural Language Instructions for Object Placement
    Zhao, Zirui
    Lee, Wee Sun
    Hsu, David
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 11546 - 11553
  • [43] Latent semantic information in maximum entropy language models for conversational speech recognition
    Deng, YG
    Khudanpur, S
    HLT-NAACL 2003: HUMAN LANGUAGE TECHNOLOGY CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE MAIN CONFERENCE, 2003, : 56 - 63
  • [44] AMR-based Semantic Parsing for the Portuguese Language
    Anchieta, Rafael Torres
    Pardo, Thiago Alexandre Salgueiro
    LINGUAMATICA, 2022, 14 (01): : 33 - 48
  • [45] Learning Structured Natural Language Representations for Semantic Parsing
    Cheng, Jianpeng
    Reddy, Siva
    Saraswat, Vijay
    Lapata, Mirella
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 44 - 55
  • [46] BertNet: Harvesting Knowledge Graphs with Arbitrary Relations from Pretrained Language Models
    Hao, Shibo
    Tan, Bowen
    Tang, Kaiwen
    Ni, Bin
    Shao, Xiyan
    Zhang, Hengzhe
    Xing, Eric P.
    Hu, Zhiting
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5000 - 5015
  • [47] Semantic grounding in models of analogy: an environmental approach
    Ramscar, M
    Yarlett, D
    COGNITIVE SCIENCE, 2003, 27 (01) : 41 - 71
  • [48] Multilingual Knowledge Graph Completion from Pretrained Language Models with Knowledge Constraints
    Song, Ran
    He, Shizhu
    Gao, Shengxiang
    Cai, Li
    Liu, Kang
    Yu, Zhengtao
    Zhao, Jun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 7709 - 7721
  • [49] Constructing Chinese taxonomy trees from understanding and generative pretrained language models
    Guo, Jianyu
    Chen, Jingnan
    Ren, Li
    Zhou, Huanlai
    Xu, Wenbo
    Jia, Haitao
    PEERJ COMPUTER SCIENCE, 2024, 10
  • [50] Pretrain-KGE: Learning Knowledge Representation from Pretrained Language Models
    Zhang, Zhiyuan
    Liu, Xiaoqian
    Zhang, Yi
    Su, Qi
    Sun, Xu
    He, Bin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 259 - 266