Understanding stance classification of BERT models: an attention-based framework

被引:0
|
作者
Carlos Abel Córdova Sáenz
Karin Becker
机构
[1] Federal University of Rio Grande do Sul (UFRGS),Institute of Informatics
来源
关键词
BERT; Interpretability; Attention; Stance classification;
D O I
暂无
中图分类号
学科分类号
摘要
BERT produces state-of-the-art solutions for many natural language processing tasks at the cost of interpretability. As works discuss the value of BERT’s attention weights to this purpose, we contribute to the field by examining this issue in the context of stance classification. We propose an interpretability framework to identify the most influential words for correctly predicting stances using BERT models. Unlike related work, we develop a broader level of interpretability focused on the overall model behaviour, aggregating tokens’ attentions into words’ attention weights that can be semantically related to the domain and proposing metrics to measure words relevance in correct predictions. We developed a broad experimental setting to analyse the premises underlying our framework regarding word attention scores and the capability concerning interpretability, adopting three case studies of stances expressed on Twitter on issues about the pandemic, and four pre-trained BERT models. We concluded that our method is not affected by the characteristics of BERT-models vocabularies, that words with high absolute attention have a higher probability of positive influence on correct classification, and that the influential words represent the domains. We observed many common words compared to a baseline method, but the words yielded by our method were considered more relevant according to a qualitative assessment.
引用
收藏
页码:419 / 451
页数:32
相关论文
共 50 条
  • [1] Understanding stance classification of BERT models: an attention-based framework
    Saenz, Carlos Abel Cordova
    Becker, Karin
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2023, 66 (1) : 419 - 451
  • [2] Assessing the use of attention weights to interpret BERT-based stance classification
    Cordova Saenz, Carlos Abel
    Becker, Karin
    [J]. 2021 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY (WI-IAT 2021), 2021, : 194 - 201
  • [3] Towards Understanding Attention-Based Speech Recognition Models
    Qin, Chu-Xiong
    Qu, Dan
    [J]. IEEE ACCESS, 2020, 8 : 24358 - 24369
  • [4] Attention-Based Model for Accurate Stance Detection
    Hamad, Omama
    Hamdi, Ali
    Shaban, Khaled
    [J]. TEXT, SPEECH, AND DIALOGUE (TSD 2022), 2022, 13502 : 212 - 224
  • [5] Few-shot relation classification by context attention-based prototypical networks with BERT
    Bei Hui
    Liang Liu
    Jia Chen
    Xue Zhou
    Yuhui Nian
    [J]. EURASIP Journal on Wireless Communications and Networking, 2020
  • [6] Few-shot relation classification by context attention-based prototypical networks with BERT
    Hui, Bei
    Liu, Liang
    Chen, Jia
    Zhou, Xue
    Nian, Yuhui
    [J]. EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2020, 2020 (01)
  • [7] Sentiment classification of microblog: A framework based on BERT and CNN with attention mechanism
    Jia, Keliang
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2022, 101
  • [8] Company Industry Classification with Neural and Attention-Based Learning Models
    Slavov, Stanislav
    Tagarev, Andrey
    Tulechki, Nikola
    Boytcheva, Svetla
    [J]. 2019 BIG DATA, KNOWLEDGE AND CONTROL SYSTEMS ENGINEERING (BDKCSE), 2019,
  • [9] On Exploring Attention-based Explanation for Transformer Models in Text Classification
    Liu, Shengzhong
    Le, Franck
    Chakraborty, Supriyo
    Abdelzaher, Tarek
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 1193 - 1203
  • [10] A BERT-Based Hybrid Short Text Classification Model Incorporating CNN and Attention-Based BiGRU
    Bao, Tong
    Ren, Ni
    Luo, Rui
    Wang, Baojia
    Shen, Gengyu
    Guo, Ting
    [J]. JOURNAL OF ORGANIZATIONAL AND END USER COMPUTING, 2021, 33 (06)