Analyzing Encoded Concepts in Transformer Language Models

被引:0
|
作者
Sajjad, Hassan [1 ]
Durrani, Nadir [1 ]
Dalvi, Fahim [1 ]
Alam, Firoj [1 ]
Khan, Abdul Rafae [2 ]
Xu, Jia [2 ]
机构
[1] Qatar Comp Res Inst, HBKU Res Complex, Ar Rayyan, Qatar
[2] Stevens Inst Technol, Sch Engn & Sci, Hoboken, NJ 07030 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a novel framework ConceptX, to analyze how latent concepts are encoded in representations learned within pre-trained language models. It uses clustering to discover the encoded concepts and explains them by aligning with a large set of human-defined concepts. Our analysis on seven transformer language models reveal interesting insights: i) the latent space within the learned representations overlap with different linguistic concepts to a varying degree, ii) the lower layers in the model are dominated by lexical concepts (e.g., affixation), whereas the core-linguistic concepts (e.g., morphological or syntactic relations) are better represented in the middle and higher layers, iii) some encoded concepts are multi-faceted and cannot be adequately explained using the existing human-defined concepts.(1)
引用
收藏
页码:3082 / 3101
页数:20
相关论文
共 50 条
  • [1] LM Transparency Tool: Interactive Tool for Analyzing Transformer Language Models
    Tufanov, Igor
    Hambardzumyan, Karen
    Ferrando, Javier
    Voita, Elena
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 3: SYSTEM DEMONSTRATIONS, 2024, : 51 - 60
  • [2] EVALUATIVE CONCEPTS ENCODED IN METAPHORICAL LANGUAGE
    Chekulai, Igor. V.
    Prokhorova, Olga N.
    Pupynina, Elena V.
    Bekh, Ekaterina F.
    REVISTA GENERO & DIREITO, 2019, 8 (06): : 478 - 491
  • [3] Analyzing Redundancy in Pretrained Transformer Models
    Dalvi, Fahim
    Sajjad, Hassan
    Durrani, Nadir
    Belinkov, Yonatan
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 4908 - 4926
  • [4] Analyzing the Structure of Attention in a Transformer Language Model
    Vig, Jesse
    Belinkov, Yonatan
    BLACKBOXNLP WORKSHOP ON ANALYZING AND INTERPRETING NEURAL NETWORKS FOR NLP AT ACL 2019, 2019, : 63 - 76
  • [5] LVCSR with Transformer Language Models
    Beck, Eugen
    Schlueter, Ralf
    Ney, Hermann
    INTERSPEECH 2020, 2020, : 1798 - 1802
  • [6] Are Structural Concepts Universal in Transformer Language Models? Towards Interpretable Cross-Lingual Generalization
    Xu, Ningyu
    Zhang, Qi
    Ye, Jingting
    Zhang, Menghan
    Huang, Xuanjing
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13951 - 13976
  • [7] Molecular language models: RNNs or transformer?
    Chen, Yangyang
    Wang, Zixu
    Zeng, Xiangxiang
    Li, Yayang
    Li, Pengyong
    Ye, Xiucai
    Sakurai, Tetsuya
    BRIEFINGS IN FUNCTIONAL GENOMICS, 2023, 22 (04) : 392 - 400
  • [8] Structural Guidance for Transformer Language Models
    Qian, Peng
    Naseem, Tahira
    Levy, Roger
    Astudillo, Ramon Fernandez
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 3735 - 3745
  • [9] Staged Training for Transformer Language Models
    Shen, Sheng
    Walsh, Pete
    Keutzer, Kurt
    Dodge, Jesse
    Peters, Matthew
    Beltagy, Iz
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [10] Analyzing the Robustness of Vision & Language Models
    Shirnin, Alexander
    Andreev, Nikita
    Potapova, Sofia
    Artemova, Ekaterina
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 2751 - 2763