共 50 条
- [2] Learning Student-Friendly Teacher Networks for Knowledge Distillation ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
- [5] Pea-KD: Parameter-efficient and accurate Knowledge Distillation on BERT PLOS ONE, 2022, 17 (02):
- [6] SFT-KD-Recon: Learning a Student-friendly Teacher for Knowledge Distillation in Magnetic Resonance Image Reconstruction MEDICAL IMAGING WITH DEEP LEARNING, VOL 227, 2023, 227 : 1423 - 1440
- [10] MiniALBERT: Model Distillation via Parameter-Efficient Recursive Transformers 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1161 - 1173