Evaluating Explanations: How Much Do Explanations from the Teacher Aid Students?

被引:23
|
作者
Pruthi, Danish [1 ]
Bansal, Rachit [2 ]
Dhingra, Bhuwan [3 ]
Soares, Livio Baldini [3 ]
Collins, Michael [3 ]
Lipton, Zachary C. [1 ]
Neubig, Graham [1 ]
Cohen, William W. [3 ]
机构
[1] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
[2] Delhi Technol Univ, New Delhi, India
[3] Google Res, Mountain View, CA USA
关键词
Automatic modeling - Model learning - Modeling architecture - Question Answering - Salient features - Student Modeling - Teacher models - Teachers' - Test time - Text classification;
D O I
10.1162/tacl_a_00465
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While many methods purport to explain predictions by highlighting salient features, what aims these explanations serve and how they ought to be evaluated often go unstated. In this work, we introduce a framework to quantify the value of explanations via the accuracy gains that they confer on a student model trained to simulate a teacher model. Crucially, the explanations are available to the student during training, but are not available at test time. Compared with prior proposals, our approach is less easily gamed, enabling principled, automatic, model-agnostic evaluation of attributions. Using our framework, we compare numerous attribution methods for text classification and question answering, and observe quantitative differences that are consistent (to a moderate to high degree) across different student model architectures and learning strategies.
引用
收藏
页码:359 / 375
页数:17
相关论文
共 50 条