共 50 条
- [34] Knowledge Distillation for Optimization of Quantized Deep Neural Networks [J]. 2020 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS), 2020, : 111 - 116
- [35] PDD: Pruning Neural Networks During Knowledge Distillation [J]. COGNITIVE COMPUTATION, 2024, : 3457 - 3467
- [36] Improving the Interpretability of Deep Neural Networks with Knowledge Distillation [J]. 2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW), 2018, : 905 - 912
- [37] Distilling Spikes: Knowledge Distillation in Spiking Neural Networks [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 4536 - 4543
- [38] Representation Learning of Knowledge Graph for Wireless Communication Networks [J]. 2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 1338 - 1343
- [39] Distilling Holistic Knowledge with Graph Neural Networks [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 10367 - 10376
- [40] Understanding the Representation Power of Graph Neural Networks in Learning Graph Topology [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32