Assessing the use of attention weights to interpret BERT-based stance classification

被引:3
|
作者
Cordova Saenz, Carlos Abel [1 ]
Becker, Karin [1 ]
机构
[1] Fed Univ Rio Grande do Sul UFRGS, Inst Informat, Porto Alegre, RS, Brazil
关键词
BERT; interpretability; stance classification; BERT attention weights;
D O I
10.1145/3486622.3493966
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
BERT models are currently state-of-the-art solutions for various tasks, including stance classification. However, these models are a black box for their users. Some proposals have leveraged the weights assigned by the internal attention mechanisms of these models for interpretability purposes. However, whether the attention weights help the interpretability of the model is still a matter of debate, with positions in favor and against. This work proposes an attention-based interpretability mechanism to identify the most influential words for stances predicted using BERT-based models. We target stances expressed in Twitter using the Portuguese language and assess the proposed mechanism using a case study regarding stances on COVID-19 vaccination in the Brazilian context. The interpretation mechanism traces tokens' attentions back to words, assigning a newly proposed metric referred to as absolute word attention. Through this metric, we assess several aspects to determine if we can find important words for the classification and with meaning for the domain. We developed a broad experimental setting that involved three datasets with tweets in Brazilian Portuguese and three BERT models with support for this language. Our results are encouraging, as we were able to identify 52-82% of words with high absolute attention contributing positively to stance classification. The interpretability mechanism proved to be helpful to understand the influence of words in the classification, and they revealed intrinsic properties of the domain and representative arguments of the stances.
引用
收藏
页码:194 / 201
页数:8
相关论文
共 50 条
  • [1] Understanding stance classification of BERT models: an attention-based framework
    Saenz, Carlos Abel Cordova
    Becker, Karin
    KNOWLEDGE AND INFORMATION SYSTEMS, 2023, 66 (1) : 419 - 451
  • [2] Understanding stance classification of BERT models: an attention-based framework
    Carlos Abel Córdova Sáenz
    Karin Becker
    Knowledge and Information Systems, 2024, 66 : 419 - 451
  • [3] BERT-Based GitHub Issue Report Classification
    Siddiq, Mohammed Latif
    Santos, Joanna C. S.
    2022 IEEE/ACM 1ST INTERNATIONAL WORKSHOP ON NATURAL LANGUAGE-BASED SOFTWARE ENGINEERING (NLBSE 2022), 2022, : 33 - 36
  • [4] A BERT-Based Hybrid Short Text Classification Model Incorporating CNN and Attention-Based BiGRU
    Bao, Tong
    Ren, Ni
    Luo, Rui
    Wang, Baojia
    Shen, Gengyu
    Guo, Ting
    JOURNAL OF ORGANIZATIONAL AND END USER COMPUTING, 2021, 33 (06)
  • [5] BAE: BERT-based Adversarial Examples for Text Classification
    Garg, Siddhant
    Ramakrishnan, Goutham
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 6174 - 6181
  • [6] A BERT-based interactive attention network for aspect sentiment analysis
    Yang, Yu-Ting
    Feng, Lin
    Dai, Lei-Chao
    Journal of Computers (Taiwan), 2021, 32 (03) : 30 - 42
  • [7] Biomedical Abstract Sentence Classification by BERT-Based Reading Comprehension
    Jiang C.-Y.
    Fan Y.-C.
    SN Computer Science, 4 (4)
  • [8] Arabic Synonym BERT-based Adversarial Examples for Text Classification
    Alshahrani, Norah
    Alshahrani, Saied
    Wali, Esma
    Matthews, Jeanna
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: STUDENT RESEARCH WORKSHOP, 2024, : 137 - 147
  • [9] Attention-Enhancing Backdoor Attacks Against BERT-based Models
    Lyu, Weimin
    Zheng, Songzhu
    Pang, Lu
    Ling, Haibin
    Chen, Chao
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 10672 - 10690
  • [10] Assessing a BERT-based model for analyzing subjectivity and classifying academic articles
    Mehmood A.
    Shahid F.
    Khan R.
    Ahmed S.
    Ibrahim M.M.
    Zheng Z.
    Multimedia Tools and Applications, 2024, 83 (42) : 90511 - 90532