共 50 条
- [11] Accelerating neural network training with distributed asynchronous and selective optimization (DASO) [J]. Journal of Big Data, 9
- [12] Distributed Deep Neural Network Training on Edge Devices [J]. SEC'19: PROCEEDINGS OF THE 4TH ACM/IEEE SYMPOSIUM ON EDGE COMPUTING, 2019, : 304 - 306
- [13] Hierarchical Distributed-Memory Multi-Leader MPI-Allreduce for Deep Learning Workloads [J]. 2018 SIXTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING WORKSHOPS (CANDARW 2018), 2018, : 216 - 222
- [14] Accelerating Allreduce With In-Network Reduction on Intel PIUMA [J]. IEEE MICRO, 2022, 42 (02) : 44 - 52
- [15] Accelerating deep neural network training for action recognition on a cluster of GPUs [J]. 2018 30TH INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE AND HIGH PERFORMANCE COMPUTING (SBAC-PAD 2018), 2018, : 298 - 305
- [17] PipePar: A Pipelined Hybrid Parallel Approach for Accelerating Distributed DNN Training [J]. PROCEEDINGS OF THE 2021 IEEE 24TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN (CSCWD), 2021, : 470 - 475
- [18] Accelerating Large-Scale Distributed Neural Network Training with SPMD Parallelism [J]. PROCEEDINGS OF THE 13TH SYMPOSIUM ON CLOUD COMPUTING, SOCC 2022, 2022, : 403 - 418
- [20] High Performance Training of Deep Neural Networks Using Pipelined Hardware Acceleration and Distributed Memory [J]. 2018 19TH INTERNATIONAL SYMPOSIUM ON QUALITY ELECTRONIC DESIGN (ISQED), 2018, : 383 - 388