On the Explainability of Natural Language Processing Deep Models

被引:32
|
作者
El Zini, Julia [1 ]
Awad, Mariette [1 ]
机构
[1] Amer Univ Beirut, Dept Elect & Comp Engn, POB 11-0236, Beirut 11072020, Lebanon
关键词
ExAI; NLP; language models; transformers; neural machine translation; transparent embedding models; explaining decisions; NEURAL-NETWORKS; GAME;
D O I
10.1145/3529755
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Despite their success, deep networks are used as black-box models with outputs that are not easily explainable during the learning and the prediction phases. This lack of interpretability is significantly limiting the adoption of such models in domains where decisions are critical such as the medical and legal fields. Recently, researchers have been interested in developing methods that help explain individual decisions and decipher the hidden representations of machine learning models in general and deep networks specifically. While there has been a recent explosion of work on Explainable Artificial Intelligence (ExAI) on deep models that operate on imagery and tabular data, textual datasets present new challenges to the ExAI community. Such challenges can be attributed to the lack of input structure in textual data, the use of word embeddings that add to the opacity of the models and the difficulty of the visualization of the inner workings of deep models when they are trained on textual data. Lately, methods have been developed to address the aforementioned challenges and present satisfactory explanations on Natural Language Processing (NLP) models. However, such methods are yet to be studied in a comprehensive framework where common challenges are properly stated and rigorous evaluation practices and metrics are proposed. Motivated to democratize ExAI methods in the NLP field, we present in this work a survey that studies model-agnostic as well as model-specific explainability methods on NLP models. Such methods can either develop inherently interpretable NLP models or operate on pre-trained models in a post hoc manner. We make this distinction and we further decompose the methods into three categories according to what they explain: (1) word embeddings (input level), (2) inner workings of NLP models (processing level), and (3) models' decisions (output level). We also detail the different evaluation approaches interpretability methods in the NLP field. Finally, we present a case-study on the well-known neural machine translation in an appendix, and we propose promising future research directions for ExAl in the NLP field.
引用
收藏
页数:31
相关论文
共 50 条
  • [21] Using deep learning and natural language processing models to detect child physical abuse
    Shahi, Niti
    Shahi, Ashwani K.
    Phillips, Ryan
    Shirek, Gabrielle
    Lindberg, Daniel M.
    Moulton, Steven L.
    [J]. JOURNAL OF PEDIATRIC SURGERY, 2021, 56 (12) : 2326 - 2332
  • [22] Using Deep Learning and Natural Language Processing Models to Detect Child Physical Abuse
    Shahi, Niti
    Shahi, Ashwani
    Phillips, Ryan
    Shirek, Gabrielle P.
    Lindberg, Daniel
    Moulton, Steven L.
    [J]. PEDIATRICS, 2021, 147 (03)
  • [23] Development of language resources for natural language processing in deep level
    Zhang, Yujie
    Kuroda, Kow
    Izumi, Emi
    Nozawa, Hajime
    [J]. Journal of the National Institute of Information and Communications Technology, 2007, 54 (03): : 53 - 61
  • [24] Dissecting word embeddings and language models in natural language processing
    Verma, Vivek Kumar
    Pandey, Mrigank
    Jain, Tarun
    Tiwari, Pradeep Kumar
    [J]. JOURNAL OF DISCRETE MATHEMATICAL SCIENCES & CRYPTOGRAPHY, 2021, 24 (05): : 1509 - 1515
  • [25] On the Reliability and Explainability of Language Models for Program Generation
    Liu, Yue
    Tantithamthavorn, Chakkrit
    Liu, Yonghui
    Li, Li
    [J]. ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2024, 33 (05)
  • [26] Exploration of biomedical knowledge for recurrent glioblastoma using natural language processing deep learning models
    Bum-Sup Jang
    Andrew J. Park
    In Ah Kim
    [J]. BMC Medical Informatics and Decision Making, 22
  • [27] Dementia Detection using Transformer-Based Deep Learning and Natural Language Processing Models
    Saltz, Ploypaphat
    Lin, Shih Yin
    Cheng, Sunny Chieh
    Si, Dong
    [J]. 2021 IEEE 9TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI 2021), 2021, : 509 - 510
  • [28] Exploration of biomedical knowledge for recurrent glioblastoma using natural language processing deep learning models
    Jang, Bum-Sup
    Park, Andrew J.
    Kim, In Ah
    [J]. BMC MEDICAL INFORMATICS AND DECISION MAKING, 2022, 22 (01)
  • [29] Special issue on deep learning for natural language processing
    Wei Wei
    Jinsong Wu
    Chunsheng Zhu
    [J]. Computing, 2020, 102 : 601 - 603
  • [30] Deep learning for natural language processing:advantages and challenges
    Hang Li
    [J]. National Science Review, 2018, 5 (01) : 24 - 26