Assessing the use of attention weights to interpret BERT-based stance classification

被引:3
|
作者
Cordova Saenz, Carlos Abel [1 ]
Becker, Karin [1 ]
机构
[1] Fed Univ Rio Grande do Sul UFRGS, Inst Informat, Porto Alegre, RS, Brazil
关键词
BERT; interpretability; stance classification; BERT attention weights;
D O I
10.1145/3486622.3493966
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
BERT models are currently state-of-the-art solutions for various tasks, including stance classification. However, these models are a black box for their users. Some proposals have leveraged the weights assigned by the internal attention mechanisms of these models for interpretability purposes. However, whether the attention weights help the interpretability of the model is still a matter of debate, with positions in favor and against. This work proposes an attention-based interpretability mechanism to identify the most influential words for stances predicted using BERT-based models. We target stances expressed in Twitter using the Portuguese language and assess the proposed mechanism using a case study regarding stances on COVID-19 vaccination in the Brazilian context. The interpretation mechanism traces tokens' attentions back to words, assigning a newly proposed metric referred to as absolute word attention. Through this metric, we assess several aspects to determine if we can find important words for the classification and with meaning for the domain. We developed a broad experimental setting that involved three datasets with tweets in Brazilian Portuguese and three BERT models with support for this language. Our results are encouraging, as we were able to identify 52-82% of words with high absolute attention contributing positively to stance classification. The interpretability mechanism proved to be helpful to understand the influence of words in the classification, and they revealed intrinsic properties of the domain and representative arguments of the stances.
引用
收藏
页码:194 / 201
页数:8
相关论文
共 50 条
  • [1] Understanding stance classification of BERT models: an attention-based framework
    Saenz, Carlos Abel Cordova
    Becker, Karin
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2023, 66 (1) : 419 - 451
  • [2] Understanding stance classification of BERT models: an attention-based framework
    Carlos Abel Córdova Sáenz
    Karin Becker
    [J]. Knowledge and Information Systems, 2024, 66 : 419 - 451
  • [3] BERT-Based GitHub Issue Report Classification
    Siddiq, Mohammed Latif
    Santos, Joanna C. S.
    [J]. 2022 IEEE/ACM 1ST INTERNATIONAL WORKSHOP ON NATURAL LANGUAGE-BASED SOFTWARE ENGINEERING (NLBSE 2022), 2022, : 33 - 36
  • [4] A BERT-Based Hybrid Short Text Classification Model Incorporating CNN and Attention-Based BiGRU
    Bao, Tong
    Ren, Ni
    Luo, Rui
    Wang, Baojia
    Shen, Gengyu
    Guo, Ting
    [J]. JOURNAL OF ORGANIZATIONAL AND END USER COMPUTING, 2021, 33 (06)
  • [5] BAE: BERT-based Adversarial Examples for Text Classification
    Garg, Siddhant
    Ramakrishnan, Goutham
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 6174 - 6181
  • [6] A BERT-based interactive attention network for aspect sentiment analysis
    Yang, Yu-Ting
    Feng, Lin
    Dai, Lei-Chao
    [J]. Journal of Computers (Taiwan), 2021, 32 (03) : 30 - 42
  • [7] Biomedical Abstract Sentence Classification by BERT-Based Reading Comprehension
    Jiang C.-Y.
    Fan Y.-C.
    [J]. SN Computer Science, 4 (4)
  • [8] Improving BERT-Based Text Classification With Auxiliary Sentence and Domain Knowledge
    Yu, Shanshan
    Su, Jindian
    Luo, Da
    [J]. IEEE ACCESS, 2019, 7 : 176600 - 176612
  • [9] Improving Bert-Based Model for Medical Text Classification with an Optimization Algorithm
    Gasmi, Karim
    [J]. ADVANCES IN COMPUTATIONAL COLLECTIVE INTELLIGENCE, ICCCI 2022, 2022, 1653 : 101 - 111
  • [10] BERT-based semi-supervised domain adaptation for disastrous classification
    Jing Wang
    Kexin Wang
    [J]. Multimedia Systems, 2022, 28 : 2237 - 2246