Analyzing Encoded Concepts in Transformer Language Models

被引:0
|
作者
Sajjad, Hassan [1 ]
Durrani, Nadir [1 ]
Dalvi, Fahim [1 ]
Alam, Firoj [1 ]
Khan, Abdul Rafae [2 ]
Xu, Jia [2 ]
机构
[1] Qatar Comp Res Inst, HBKU Res Complex, Ar Rayyan, Qatar
[2] Stevens Inst Technol, Sch Engn & Sci, Hoboken, NJ 07030 USA
来源
NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES | 2022年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a novel framework ConceptX, to analyze how latent concepts are encoded in representations learned within pre-trained language models. It uses clustering to discover the encoded concepts and explains them by aligning with a large set of human-defined concepts. Our analysis on seven transformer language models reveal interesting insights: i) the latent space within the learned representations overlap with different linguistic concepts to a varying degree, ii) the lower layers in the model are dominated by lexical concepts (e.g., affixation), whereas the core-linguistic concepts (e.g., morphological or syntactic relations) are better represented in the middle and higher layers, iii) some encoded concepts are multi-faceted and cannot be adequately explained using the existing human-defined concepts.(1)
引用
收藏
页码:3082 / 3101
页数:20
相关论文
共 50 条
  • [21] Teaching Structured Vision & Language Concepts to Vision & Language Models
    Doveh, Sivan
    Arbelle, Assaf
    Harary, Sivan
    Schwartz, Eli
    Herzig, Roei
    Giryes, Raja
    Feris, Rogerio
    Panda, Rameswar
    Ullman, Shimon
    Karlinsky, Leonid
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 2657 - 2668
  • [22] Exemplar models in the study of natural language concepts
    Storms, G
    PSYCHOLOGY OF LEARNING AND MOTIVATION: ADVANCES IN RESEARCH AND THEORY, VOL 45, 2004, 45 : 1 - 39
  • [23] Can Language Models Understand Physical Concepts?
    Li, Lei
    Xu, Jingjing
    Dong, Qingxiu
    Zheng, Ce
    Sun, Xu
    Kong, Lingpeng
    Liu, Qi
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 11843 - 11861
  • [24] Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
    Sartran, Laurent
    Barrett, Samuel
    Kuncoro, Adhiguna
    Stanojevic, Milos
    Blunsom, Phil
    Dyer, Chris
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 1423 - 1439
  • [25] Propositional Reasoning via Neural Transformer Language Models
    Tomasic, Anthony
    Romero, Oscar J.
    Zimmerman, John
    Steinfeld, Aaron
    NEURAL-SYMBOLIC LEARNING AND REASONING, NESY 2022, 2022, : 104 - 119
  • [26] Can Transformer Language Models Predict Psychometric Properties?
    Laverghetta, Antonio, Jr.
    Nighojkar, Animesh
    Mirzakhalov, Jamshidbek
    Licato, John
    10TH CONFERENCE ON LEXICAL AND COMPUTATIONAL SEMANTICS (SEM 2021), 2021, : 12 - 25
  • [27] Improved Hybrid Streaming ASR with Transformer Language Models
    Baquero-Arnal, Pau
    Jorge, Javier
    Gimenez, Adria
    Albert Silvestre-Cerda, Joan
    Iranzo-Sanchez, Javier
    Sanchis, Albert
    Civera, Jorge
    Juan, Alfons
    INTERSPEECH 2020, 2020, : 2127 - 2131
  • [28] The concepts of language and the language of concepts
    Cimino, JJ
    METHODS OF INFORMATION IN MEDICINE, 1998, 37 (4-5) : 311 - 311
  • [29] Inspecting the concept knowledge graph encoded by modern language models
    Aspillaga, Carlos
    Mendoza, Marcelo
    Soto, Alvaro
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 2984 - 3000
  • [30] Analyzing Declarative Deployment Code with Large Language Models
    Lanciano, Giacomo
    Stein, Manuel
    Hilt, Volker
    Cucinotta, Tommaso
    PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND SERVICES SCIENCE, CLOSER 2023, 2023, : 289 - 296