Causal Distillation for Language Models

被引:0
|
作者
Wu, Zhengxuan [1 ]
Geiger, Atticus [1 ]
Rozner, Joshua [1 ]
Kreiss, Elisa [1 ]
Lu, Hanson [1 ]
Icard, Thomas [1 ]
Potts, Christopher [1 ]
Goodman, Noah [1 ]
机构
[1] Stanford Univ, Stanford, CA 94305 USA
关键词
EXPLANATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Distillation efforts have led to language models that are more compact and efficient without serious drops in performance. The standard approach to distillation trains a student model against two objectives: a task-specific objective (e.g., language modeling) and an imitation objective that encourages the hidden states of the student model to be similar to those of the larger teacher model. In this paper, we show that it is beneficial to augment distillation with a third objective that encourages the student to imitate the causal dynamics of the teacher through a distillation interchange intervention training objective (DIITO). DIITO pushes the student model to become a causal abstraction of the teacher model - a faithful model with simpler causal structure. DIITO is fully differentiable, easily implemented, and combines flexibly with other objectives. Compared against standard distillation with the same setting, DIITO results in lower perplexity on the WikiText-103M corpus (masked language modeling) and marked improvements on the GLUE benchmark (natural language understanding), SQuAD (question answering), and CoNLL-2003 (named entity recognition).
引用
收藏
页码:4288 / 4295
页数:8
相关论文
共 50 条
  • [1] Neurobiological Causal Models of Language Processing
    Fitz, Hartmut
    Hagoort, Peter
    Petersson, Karl Magnus
    [J]. NEUROBIOLOGY OF LANGUAGE, 2024, 5 (01): : 225 - 247
  • [2] Causal Dataset Discovery with Large Language Models
    Liu, Junfei
    Sun, Shaotong
    Nargesian, Fatemeh
    [J]. WORKSHOP ON HUMAN-IN-THE-LOOP DATA ANALYTICS, HILDA 2024, 2024,
  • [3] CLADDER: Assessing Causal Reasoning in Language Models
    Jin, Zhijing
    Chen, Yuen
    Leeb, Felix
    Gresele, Luigi
    Kamal, Ojasv
    Lyu, Zhiheng
    Blin, Kevin
    Gonzalez, Fernando
    Kleiman-Weiner, Max
    Sachan, Mrinmaya
    Schoelkopf, Bernhard
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [4] Symbolic Knowledge Distillation: from General Language Models to Commonsense Models
    West, Peter
    Bhagavatula, Chandra
    Hessel, Jack
    Hwang, Jena D.
    Jiang, Liwei
    Le Bras, Ronan
    Lu, Ximing
    Welleck, Sean
    Choi, Yejin
    [J]. NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 4602 - 4625
  • [5] Dynamic Knowledge Distillation for Pre-trained Language Models
    Li, Lei
    Lin, Yankai
    Ren, Shuhuai
    Li, Peng
    Zhou, Jie
    Sun, Xu
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 379 - 389
  • [6] A Causal Framework to Quantify the Robustness of Mathematical Reasoning with Language Models
    Stolfo, Alessandro
    Jin, Zhijing
    Shridhar, Kumar
    Scholkopf, Bernhard
    Sachan, Mrinmaya
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 545 - 561
  • [7] Causal Analysis of Syntactic Agreement Mechanisms in Neural Language Models
    Finlayson, Matthew
    Mueller, Aaron
    Gehrmann, Sebastian
    Shieber, Stuart
    Linzen, Tal
    Belinkov, Yonatan
    [J]. 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 1828 - 1843
  • [8] CausaLM: Causal Model Explanation Through Counterfactual Language Models
    Feder, Amir
    Oved, Nadav
    Shalit, Uri
    Reichart, Roi
    [J]. COMPUTATIONAL LINGUISTICS, 2021, 47 (02) : 333 - 386
  • [9] Passive learning of active causal strategies in agents and language models
    Lampinen, Andrew K.
    Chan, Stephanie C. Y.
    Dasgupta, Ishita
    Nam, Andrew J.
    Wang, Jane X.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [10] Scalable Syntax-Aware Language Models Using Knowledge Distillation
    Kuncoro, Adhiguna
    Dyer, Chris
    Rimell, Laura
    Clark, Stephen
    Blunsom, Phil
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3472 - 3484