Explaining pretrained language models' understanding of linguistic structures using construction grammar

被引:0
|
作者
Weissweiler, Leonie [1 ,2 ]
Hofmann, Valentin [1 ,3 ]
Koeksal, Abdullatif [1 ,2 ]
Schuetze, Hinrich [1 ,2 ]
机构
[1] Ludwig Maximilians Univ Munchen, Ctr Informat & Language Proc, Munich, Germany
[2] Munich Ctr Machine Learning, Munich, Germany
[3] Univ Oxford, Fac Linguist, Oxford, England
来源
基金
欧洲研究理事会;
关键词
NLP; probing; construction grammar; computational linguistics; large language models; COMPARATIVE CORRELATIVES;
D O I
10.3389/frai.2023.1225791
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Construction Grammar (CxG) is a paradigm from cognitive linguistics emphasizing the connection between syntax and semantics. Rather than rules that operate on lexical items, it posits constructions as the central building blocks of language, i.e., linguistic units of different granularity that combine syntax and semantics. As a first step toward assessing the compatibility of CxG with the syntactic and semantic knowledge demonstrated by state-of-the-art pretrained language models (PLMs), we present an investigation of their capability to classify and understand one of the most commonly studied constructions, the English comparative correlative (CC). We conduct experiments examining the classification accuracy of a syntactic probe on the one hand and the models' behavior in a semantic application task on the other, with BERT, RoBERTa, and DeBERTa as the example PLMs. Our results show that all three investigated PLMs, as well as OPT, are able to recognize the structure of the CC but fail to use its meaning. While human-like performance of PLMs on many NLP tasks has been alleged, this indicates that PLMs still suffer from substantial shortcomings in central domains of linguistic knowledge.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Rethinking the Construction of Effective Metrics for Understanding the Mechanisms of Pretrained Language Models
    Li, You
    Yin, Jinhui
    Lin, Yuming
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13399 - 13412
  • [2] Language and creativity: a Construction Grammar approach to linguistic creativity
    Hoffmann, Thomas
    LINGUISTICS VANGUARD, 2019, 5 (01):
  • [3] Data Augmentation for Spoken Language Understanding via Pretrained Language Models
    Peng, Baolin
    Zhu, Chenguang
    Zeng, Michael
    Gao, Jianfeng
    INTERSPEECH 2021, 2021, : 1219 - 1223
  • [4] Probing for Predicate Argument Structures in Pretrained Language Models
    Conia, Simone
    Navigli, Roberto
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4622 - 4632
  • [5] The Heuristic Core: Understanding Subnetwork Generalization in Pretrained Language Models
    Bhaskar, Adithya
    Chen, Danqi
    Friedman, Dan
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 14351 - 14368
  • [6] Augmenting Slot Values and Contexts for Spoken Language Understanding with Pretrained Models
    Lin, Haitao
    Xiang, Lu
    Zhou, Yu
    Zhang, Jiajun
    Zong, Chengqing
    INTERSPEECH 2021, 2021, : 4703 - 4707
  • [7] LEVERAGING ACOUSTIC AND LINGUISTIC EMBEDDINGS FROM PRETRAINED SPEECH AND LANGUAGE MODELS FOR INTENT CLASSIFICATION
    Sharma, Bidisha
    Madhavi, Maulik
    Li, Haizhou
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7498 - 7502
  • [8] Cross-Lingual Information Retrieval from Multilingual Construction Documents Using Pretrained Language Models
    Kim, Jungyeon
    Chung, Sehwan
    Chi, Seokho
    JOURNAL OF CONSTRUCTION ENGINEERING AND MANAGEMENT, 2024, 150 (06)
  • [9] Understanding of Kazakh language with using of link grammar
    Zhumanov, Zh.
    6TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND INTELLIGENT SYSTEMS, AND THE 13TH INTERNATIONAL SYMPOSIUM ON ADVANCED INTELLIGENT SYSTEMS, 2012, : 1085 - 1088
  • [10] Constructing Chinese taxonomy trees from understanding and generative pretrained language models
    Guo, Jianyu
    Chen, Jingnan
    Ren, Li
    Zhou, Huanlai
    Xu, Wenbo
    Jia, Haitao
    PEERJ COMPUTER SCIENCE, 2024, 10