共 50 条
- [23] Large-scale Multi-modal Pre-trained Models: A Comprehensive Survey Machine Intelligence Research, 2023, 20 : 447 - 482
- [24] FASTERMOE: Modeling and Optimizing Training of Large-Scale Dynamic Pre-Trained Models PPOPP'22: PROCEEDINGS OF THE 27TH ACM SIGPLAN SYMPOSIUM ON PRINCIPLES AND PRACTICE OF PARALLEL PROGRAMMING, 2022, : 120 - 134
- [25] Alternating Recurrent Dialog Model with Large-scale Pre-trained Language Models 16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1292 - 1301
- [26] PPT: Pre-trained Prompt Tuning for Few-shot Learning PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 8410 - 8423
- [27] Large-Scale Relation Learning for Question Answering over Knowledge Bases with Pre-trained Language Models 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3653 - 3660
- [28] Hadamard Adapter: An Extreme Parameter-Efficient Adapter Tuning Method for Pre-trained Language Models PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 276 - 285
- [30] An Empirical Study of Parameter-Efficient Fine-Tuning Methods for Pre-trained Code Models 2023 38TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE, 2023, : 397 - 408