共 50 条
- [44] SSS: Editing Factual Knowledge in Language Models towards Semantic Sparse Space FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5559 - 5570
- [45] On Learning Mixture Models with Sparse Parameters INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
- [46] ReAugKD: Retrieval-Augmented Knowledge Distillation For Pre-trained Language Models 61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 1128 - 1136
- [47] DistillSeq: A Framework for Safety Alignment Testing in Large Language Models using Knowledge Distillation PROCEEDINGS OF THE 33RD ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2024, 2024, : 578 - 589
- [49] Beyond Distillation: Task-level Mixture-of-Experts for Efficient Inference FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 3577 - 3599
- [50] On the Benefits of Learning to Route in Mixture-of-Experts Models 2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 9376 - 9396