Analyzing Encoded Concepts in Transformer Language Models

被引:0
|
作者
Sajjad, Hassan [1 ]
Durrani, Nadir [1 ]
Dalvi, Fahim [1 ]
Alam, Firoj [1 ]
Khan, Abdul Rafae [2 ]
Xu, Jia [2 ]
机构
[1] Qatar Comp Res Inst, HBKU Res Complex, Ar Rayyan, Qatar
[2] Stevens Inst Technol, Sch Engn & Sci, Hoboken, NJ 07030 USA
来源
NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES | 2022年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a novel framework ConceptX, to analyze how latent concepts are encoded in representations learned within pre-trained language models. It uses clustering to discover the encoded concepts and explains them by aligning with a large set of human-defined concepts. Our analysis on seven transformer language models reveal interesting insights: i) the latent space within the learned representations overlap with different linguistic concepts to a varying degree, ii) the lower layers in the model are dominated by lexical concepts (e.g., affixation), whereas the core-linguistic concepts (e.g., morphological or syntactic relations) are better represented in the middle and higher layers, iii) some encoded concepts are multi-faceted and cannot be adequately explained using the existing human-defined concepts.(1)
引用
收藏
页码:3082 / 3101
页数:20
相关论文
共 50 条
  • [31] State gradients for analyzing memory in LSTM language models
    Verwimp, Lyan
    Van Hamme, Hugo
    Wambacq, Patrick
    COMPUTER SPEECH AND LANGUAGE, 2020, 61
  • [32] Analyzing Leakage of Personally Identifiable Information in Language Models
    Lukas, Nils
    Salem, Ahmed
    Sim, Robert
    Tople, Shruti
    Wutschitz, Lukas
    Zanella-Beguelin, Santiago
    2023 IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP, 2023, : 346 - 363
  • [33] Analyzing the Efficacy of Large Language Models: A Comparative Study
    Khetarpaul, Sonia
    Sharma, Dolly
    Sinha, Shreya
    Nagpal, Aryan
    Narang, Aarush
    DATABASE AND EXPERT SYSTEMS APPLICATIONS, PT I, DEXA 2024, 2024, 14910 : 215 - 221
  • [34] Analyzing Information Leakage of Updates to Natural Language Models
    Zanella-Beguelin, Santiago
    Wutschitz, Lukas
    Tople, Shruti
    Ruhle, Victor
    Paverd, Andrew
    Ohrimenko, Olga
    Kopf, Boris
    Brockschmidt, Marc
    CCS '20: PROCEEDINGS OF THE 2020 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2020, : 363 - 375
  • [35] Analyzing And Editing Inner Mechanisms of Backdoored Language Models
    Lamparth, Max
    Reuel, Anka
    PROCEEDINGS OF THE 2024 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, ACM FACCT 2024, 2024, : 2362 - 2373
  • [36] Locating and Extracting Relational Concepts in Large Language Models
    Wang, Zijian
    Whyte, Britney
    Xu, Chang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 4818 - 4832
  • [37] Probabilistic generative transformer language models for generative design of molecules
    Wei, Lai
    Fu, Nihang
    Song, Yuqi
    Wang, Qian
    Hu, Jianjun
    JOURNAL OF CHEMINFORMATICS, 2023, 15 (01)
  • [38] MIXED PRECISION QUANTIZATION OF TRANSFORMER LANGUAGE MODELS FOR SPEECH RECOGNITION
    Xu, Junhao
    Hu, Shoukang
    Yu, Jianwei
    Liu, Xunying
    Meng, Helen
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7383 - 7387
  • [39] Transformer-Based Language Models for Software Vulnerability Detection
    Thapa, Chandra
    Jang, Seung Ick
    Ahmed, Muhammad Ejaz
    Camtepe, Seyit
    Pieprzyk, Josef
    Nepal, Surya
    PROCEEDINGS OF THE 38TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE, ACSAC 2022, 2022, : 481 - 496
  • [40] ON THE PROTO-ROLE PROPERTIES INFERRED BY TRANSFORMER LANGUAGE MODELS
    Proietti, Mattia
    Lebani, Gianluca e.
    Lenci, Alessandro
    LINGUE E LINGUAGGIO, 2024, 23 (01) : 111 - 140