共 50 条
- [41] Distributed Framework for Accelerating Training of Deep Learning Models through Prioritization 2021 IEEE INTERNATIONAL CONFERENCE ON CLOUD ENGINEERING, IC2E 2021, 2021, : 201 - 209
- [42] Leader Stochastic Gradient Descent for Distributed Training of Deep Learning Models ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
- [43] Phase-Change Memory Models for Deep Learning Training and Inference 2019 26TH IEEE INTERNATIONAL CONFERENCE ON ELECTRONICS, CIRCUITS AND SYSTEMS (ICECS), 2019, : 727 - 730
- [44] Exploration of the Influence on Training Deep Learning Models by Watermarked Image Dataset 19TH IEEE INTERNATIONAL SYMPOSIUM ON PARALLEL AND DISTRIBUTED PROCESSING WITH APPLICATIONS (ISPA/BDCLOUD/SOCIALCOM/SUSTAINCOM 2021), 2021, : 421 - 428
- [47] Efficient Training of Deep Learning Models Through Improved Adaptive Sampling PATTERN RECOGNITION (MCPR 2021), 2021, 12725 : 141 - 152
- [49] Training confounder-free deep learning models for medical applications Nature Communications, 11