共 50 条
- [2] Accelerating distributed deep neural network training with pipelined MPI allreduce [J]. CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2021, 24 (04): : 3797 - 3813
- [3] Accelerating distributed deep neural network training with pipelined MPI allreduce [J]. Cluster Computing, 2021, 24 : 3797 - 3813
- [4] FloatPIM: In-Memory Acceleration of Deep Neural Network Training with High Precision [J]. PROCEEDINGS OF THE 2019 46TH INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA '19), 2019, : 802 - 815
- [7] A Pipelined Energy-efficient Hardware Accelaration for Deep Convolutional Neural Networks [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON DESIGN & TEST OF INTEGRATED MICRO & NANO-SYSTEMS (DTS), 2019,
- [9] Efficient Hardware Acceleration for Approximate Inference of Bitwise Deep Neural Networks [J]. 2017 CONFERENCE ON DESIGN AND ARCHITECTURES FOR SIGNAL AND IMAGE PROCESSING (DASIP), 2017,
- [10] Soft Memory Box: A Virtual Shared Memory Framework for Fast Deep Neural Network Training in Distributed High Performance Computing [J]. IEEE ACCESS, 2018, 6 : 26493 - 26504