Analyzing Encoded Concepts in Transformer Language Models

被引:0
|
作者
Sajjad, Hassan [1 ]
Durrani, Nadir [1 ]
Dalvi, Fahim [1 ]
Alam, Firoj [1 ]
Khan, Abdul Rafae [2 ]
Xu, Jia [2 ]
机构
[1] Qatar Comp Res Inst, HBKU Res Complex, Ar Rayyan, Qatar
[2] Stevens Inst Technol, Sch Engn & Sci, Hoboken, NJ 07030 USA
来源
NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES | 2022年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a novel framework ConceptX, to analyze how latent concepts are encoded in representations learned within pre-trained language models. It uses clustering to discover the encoded concepts and explains them by aligning with a large set of human-defined concepts. Our analysis on seven transformer language models reveal interesting insights: i) the latent space within the learned representations overlap with different linguistic concepts to a varying degree, ii) the lower layers in the model are dominated by lexical concepts (e.g., affixation), whereas the core-linguistic concepts (e.g., morphological or syntactic relations) are better represented in the middle and higher layers, iii) some encoded concepts are multi-faceted and cannot be adequately explained using the existing human-defined concepts.(1)
引用
收藏
页码:3082 / 3101
页数:20
相关论文
共 50 条
  • [41] Ouroboros: On Accelerating Training of Transformer-Based Language Models
    Yang, Qian
    Huo, Zhouyuan
    Wang, Wenlin
    Huang, Heng
    Carin, Lawrence
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [42] Transformer Language Models Handle Word Frequency in Prediction Head
    Kobayashi, Goro
    Kuribayashi, Tatsuki
    Yokoi, Sho
    Inui, Kentaro
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 4523 - 4535
  • [43] A Comparison of Transformer-Based Language Models on NLP Benchmarks
    Greco, Candida Maria
    Tagarelli, Andrea
    Zumpano, Ester
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2022), 2022, 13286 : 490 - 501
  • [44] Pushdown Layers: Encoding Recursive Structure in Transformer Language Models
    Murty, Shikhar
    Sharma, Pratyusha
    Andreas, Jacob
    Manning, Christopher D.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3233 - 3247
  • [45] Applications of transformer-based language models in bioinformatics: a survey
    Zhang, Shuang
    Fan, Rui
    Liu, Yuti
    Chen, Shuang
    Liu, Qiao
    Zeng, Wanwen
    NEURO-ONCOLOGY ADVANCES, 2023, 5 (01)
  • [46] TAG: Gradient Attack on Transformer-based Language Models
    Deng, Jieren
    Wang, Yijue
    Li, Ji
    Wang, Chenghong
    Shang, Chao
    Liu, Hang
    Rajasekaran, Sanguthevar
    Ding, Caiwen
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 3600 - 3610
  • [47] The Unreasonable Effectiveness of Transformer Language Models in Grammatical Error Correction
    Alikaniotis, Dimitrios
    Raheja, Vipul
    INNOVATIVE USE OF NLP FOR BUILDING EDUCATIONAL APPLICATIONS, 2019, : 127 - 133
  • [48] Probabilistic generative transformer language models for generative design of molecules
    Lai Wei
    Nihang Fu
    Yuqi Song
    Qian Wang
    Jianjun Hu
    Journal of Cheminformatics, 15
  • [49] A comparative analysis of transformer based models for figurative language classification
    Junaid, Taha
    Sumathi, D.
    Sasikumar, A. N.
    Suthir, S.
    Manikandan, J.
    Khilar, Rashmita
    Kuppusamy, P. G.
    Raju, M. Janardhana
    COMPUTERS & ELECTRICAL ENGINEERING, 2022, 101
  • [50] Ecco: An Open Source Library for the Explainability of Transformer Language Models
    Alammar, J.
    ACL-IJCNLP 2021: THE JOINT CONFERENCE OF THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING: PROCEEDINGS OF THE SYSTEM DEMONSTRATIONS, 2021, : 249 - 257