Probing Pretrained Language Models for Semantic Attributes and their Values

被引:0
|
作者
Beloucif, Meriem [1 ]
Biemann, Chris [1 ]
机构
[1] Univ Hamburg, Language Technol Grp, Dept Informat, MIN Fac, Hamburg, Germany
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pretrained Language Models (PTLMs) yield state-of-the-art performance on many Natural Language Processing tasks, including syntax, semantics and commonsense reasoning. In this paper, we focus on identifying to what extent do PTLMs capture semantic attributes and their values, e.g. the relation between rich and high net worth. We use PTLMs to predict masked tokens using patterns and lists of items from Wikidata in order to verify how likely PTLMs encode semantic attributes along with their values. Such inferences based on semantics are intuitive for us humans as part of our language understanding. Since PTLMs are trained on large amount of Wikipedia data, we would assume that they can generate similar predictions. However, our findings reveal that PTLMs perform still much worse than humans on this task. We show an analysis which explains how to exploit our methodology to integrate better context and semantics into PTLMs using knowledge bases.
引用
收藏
页码:2554 / 2559
页数:6
相关论文
共 50 条
  • [1] Discourse Probing of Pretrained Language Models
    Koto, Fajri
    Lau, Jey Han
    Baldwin, Timothy
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 3849 - 3864
  • [2] Probing Pretrained Language Models with Hierarchy Properties
    Lovon-Melgarejo, Jesus
    Moreno, Jose G.
    Besancon, Romaric
    Ferret, Olivier
    Tamine, Lynda
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT II, 2024, 14609 : 126 - 142
  • [3] Probing Pretrained Language Models for Lexical Semantics
    Vulie, Ivan
    Ponti, Edoardo M.
    Litschko, Robert
    Glava, Goran
    Korhonen, Anna
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 7222 - 7240
  • [4] Debiasing Pretrained Generative Models by Uniformly Sampling Semantic Attributes
    Gerych, Walter
    Hickey, Kevin
    Buquicchio, Luke
    Chandrasekaran, Kavin
    Alajaji, Abdulaziz
    Rundensteiner, Elke
    Agu, Emmanuel
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Probing for Predicate Argument Structures in Pretrained Language Models
    Conia, Simone
    Navigli, Roberto
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4622 - 4632
  • [6] Awakening Latent Grounding from Pretrained Language Models for Semantic Parsing
    Liu, Qian
    Yang, Dejian
    Zhang, Jiahui
    Guo, Jiaqi
    Zhou, Bin
    Lou, Jian-Guang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1174 - 1189
  • [7] Augmenting Slot Values and Contexts for Spoken Language Understanding with Pretrained Models
    Lin, Haitao
    Xiang, Lu
    Zhou, Yu
    Zhang, Jiajun
    Zong, Chengqing
    INTERSPEECH 2021, 2021, : 4703 - 4707
  • [8] A Survey of Pretrained Language Models
    Sun, Kaili
    Luo, Xudong
    Luo, Michael Y.
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT II, 2022, 13369 : 442 - 456
  • [9] DLAMA: A Framework for Curating Culturally Diverse Facts for Probing the Knowledge of Pretrained Language Models
    Keleg, Amr
    Magdy, Walid
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 6245 - 6266
  • [10] Probing Natural Language Inference Models through Semantic Fragments
    Richardson, Kyle
    Hu, Hai
    Moss, Lawrence S.
    Sabharwal, Ashish
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 8713 - 8721