共 50 条
- [1] Accelerating Transformer Inference for Translation via Parallel Decoding [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 12336 - 12355
- [2] Chemical transformer compression for accelerating both training and inference of molecular modeling [J]. MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2022, 3 (04):
- [3] Preventing Denial-of-request Inference Attacks in Location-sharing Services [J]. 2014 SEVENTH INTERNATIONAL CONFERENCE ON MOBILE COMPUTING AND UBIQUITOUS NETWORKING (ICMU), 2014, : 50 - 55
- [5] Efficient type inference for record concatenation and subtyping [J]. 17TH ANNUAL IEEE SYMPOSIUM ON LOGIC IN COMPUTER SCIENCE, PROCEEDINGS, 2002, : 125 - 136
- [9] Unleashing the Potential of PIM: Accelerating Large Batched Inference of Transformer-Based Generative Models [J]. 2024 IEEE INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE COMPUTER ARCHITECTURE, HPCA 2024, 2024, : 614 - 614
- [10] TYPE INFERENCE FOR RECORD CONCATENATION AND MULTIPLE INHERITANCE [J]. FOURTH ANNUAL SYMPOSIUM ON LOGIC IN COMPUTER SCIENCE, 1989, : 92 - 97