Probing for Bridging Inference in Transformer Language Models

被引:0
|
作者
Pandit, Onkar [1 ]
Hou, Yufang [2 ]
机构
[1] Univ Lille, CNRS, Cent Lille, INRIA Lille,UMR 9189,CRIStAL, F-59000 Lille, France
[2] IBM Res Europe, Dublin, Ireland
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We probe pre-trained transformer language models for bridging inference. We first investigate individual attention heads in BERT and observe that attention heads at higher layers prominently focus on bridging relations incomparison with the lower and middle layers, also, few specific attention heads concentrate consistently on bridging. More importantly, we consider language models as a whole in our second approach where bridging anaphora resolution is formulated as a masked token prediction task (Of-Cloze test). Our formulation produces optimistic results without any fine-tuning, which indicates that pre-trained language models substantially capture bridging inference. Our further investigation shows that the distance between anaphor-antecedent and the context provided to language models play an important role in the inference.
引用
收藏
页码:4153 / 4163
页数:11
相关论文
共 50 条
  • [21] DeepSpeed-Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale
    Aminabadi, Reza Yazdani
    Rajbhandari, Samyam
    Awan, Ammar Ahmad
    Li, Cheng
    Li, Du
    Zheng, Elton
    Ruwase, Olatunji
    Smith, Shaden
    Zhang, Minjia
    Rasley, Jeff
    He, Yuxiong
    SC22: INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2022,
  • [22] What Do Language Models Hear? Probing for Auditory Representations in Language Models
    Ngo, Jerry
    Kim, Yoon
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 5435 - 5448
  • [23] When Language Models Fall in Love: Animacy Processing in Transformer Language Models
    Hanna, Michael
    Belinkov, Yonatan
    Pezzelle, Sandro
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12120 - 12135
  • [24] Probing Pretrained Language Models for Lexical Semantics
    Vulie, Ivan
    Ponti, Edoardo M.
    Litschko, Robert
    Glava, Goran
    Korhonen, Anna
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 7222 - 7240
  • [25] Probing Pretrained Language Models with Hierarchy Properties
    Lovon-Melgarejo, Jesus
    Moreno, Jose G.
    Besancon, Romaric
    Ferret, Olivier
    Tamine, Lynda
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT II, 2024, 14609 : 126 - 142
  • [26] Developmental Negation Processing in Transformer Language Models
    Laverghetta, Antonio, Jr.
    Licato, John
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): (SHORT PAPERS), VOL 2, 2022, : 545 - 551
  • [27] BAYESIAN TRANSFORMER LANGUAGE MODELS FOR SPEECH RECOGNITION
    Xue, Boyang
    Yu, Jianwei
    Xu, Junhao
    Liu, Shansong
    Hu, Shoukang
    Ye, Zi
    Geng, Mengzhe
    Liu, Xunying
    Meng, Helen
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7378 - 7382
  • [28] Analyzing Encoded Concepts in Transformer Language Models
    Sajjad, Hassan
    Durrani, Nadir
    Dalvi, Fahim
    Alam, Firoj
    Khan, Abdul Rafae
    Xu, Jia
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 3082 - 3101
  • [29] Symbolic Semantic Memory in Transformer Language Models
    Morain, Robert
    Vargas, Kenneth
    Ventura, Dan
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 992 - 998
  • [30] IgboBERT Models: Building and Training Transformer Models for the Igbo Language
    Chukwuneke, Chiamaka
    Ezeani, Ignatius
    Rayson, Paul
    El-Haj, Mahmoud
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 5114 - 5122