共 50 条
- [4] Hardware-friendly Deep Learning by Network Quantization and Binarization [J]. PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4911 - 4912
- [5] CLIP-Q: Deep Network Compression Learning by In-Parallel Pruning-Quantization [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7873 - 7882
- [6] Octave Deep Compression: In-Parallel Pruning-Quantization on Different Frequencies [J]. 2021 IEEE 22ND INTERNATIONAL CONFERENCE ON INFORMATION REUSE AND INTEGRATION FOR DATA SCIENCE (IRI 2021), 2021, : 184 - 192
- [7] Hardware-Friendly Acceleration for Deep Neural Networks with Micro-Structured Compression [J]. 2022 IEEE 30TH INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM 2022), 2022, : 229 - 229
- [8] OPQ: Compressing Deep Neural Networks with One-shot Pruning-Quantization [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 7780 - 7788
- [9] Float-Fix: An Efficient and Hardware-Friendly Data Type for Deep Neural Network [J]. International Journal of Parallel Programming, 2019, 47 : 345 - 359
- [10] Hardware-friendly compression and hardware acceleration for transformer: A survey [J]. ELECTRONIC RESEARCH ARCHIVE, 2022, 30 (10): : 3755 - 3785