Deep Learning Model Compression With Rank Reduction in Tensor Decomposition

被引:0
|
作者
Dai, Wei [1 ,2 ]
Fan, Jicong [1 ,3 ]
Miao, Yiming [1 ,2 ]
Hwang, Kai [1 ,2 ]
机构
[1] Chinese Univ Hong Kong, Sch Data Sci, Shenzhen 518172, Peoples R China
[2] Shenzhen Inst Artificial Intelligence & Robot Soc, Shenzhen 518129, Peoples R China
[3] Shenzhen Res Inst Big Data, Shenzhen 518172, Peoples R China
基金
中国国家自然科学基金;
关键词
Tensors; Training; Matrix decomposition; Image coding; Computational modeling; Adaptation models; Deep learning; Deep learning (DL); low-rank decomposition; model compression; rank reduction (RR); NEURAL-NETWORKS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large neural network models are hard to deploy on lightweight edge devices demanding large network bandwidth. In this article, we propose a novel deep learning (DL) model compression method. Specifically, we present a dual-model training strategy with an iterative and adaptive rank reduction (RR) in tensor decomposition. Our method regularizes the DL models while preserving model accuracy. With adaptive RR, the hyperparameter search space is significantly reduced. We provide a theoretical analysis of the convergence and complexity of the proposed method. Testing our method for the LeNet, VGG, ResNet, EfficientNet, and RevCol over MNIST, CIFAR-10/100, and ImageNet datasets, our method outperforms the baseline compression methods in both model compression and accuracy preservation. The experimental results validate our theoretical findings. For the VGG-16 on CIFAR-10 dataset, our compressed model has shown a 0.88% accuracy gain with 10.41 times storage reduction and 6.29 times speedup. For the ResNet-50 on ImageNet dataset, our compressed model results in 2.36 times storage reduction and 2.17 times speedup. In federated learning (FL) applications, our scheme reduces 13.96 times the communication overhead. In summary, our compressed DL method can improve the image understanding and pattern recognition processes significantly.
引用
收藏
页码:1315 / 1328
页数:14
相关论文
共 50 条
  • [31] Robust Low-Rank and Sparse Tensor Decomposition for Low-Rank Tensor Completion
    Shi, Yuqing
    Du, Shiqiang
    Wang, Weilan
    PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, : 7138 - 7143
  • [32] Sparse and Low-Rank Tensor Decomposition
    Shah, Parikshit
    Rao, Nikhil
    Tang, Gongguo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [33] Practical Low-Rank Communication Compression in Decentralized Deep Learning
    Vogels, Thijs
    Karimireddy, Sai Praneeth
    Jaggi, Martin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [34] Active thermography based on tensor rank decomposition
    Spaeth, Peter W.
    Takunju, Peter L.
    Zalameda, Joseph N.
    THERMOSENSE: THERMAL INFRARED APPLICATIONS XLIV, 2022, 12109
  • [35] Model Reduction and Simulation of Nonlinear Circuits via Tensor Decomposition
    Liu, Haotian
    Daniel, Luca
    Wong, Ngai
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2015, 34 (07) : 1059 - 1069
  • [36] Tensor Decomposition Dictionary Learning With Low-Rank Regularization for Seismic Data Denoising
    Liu, Lina
    Liu, Zhao
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2025, 22
  • [37] Tensor decomposition for dimension reduction
    Cheng, Yu-Hsiang
    Huang, Tzee-Ming
    Huang, Su-Yun
    WILEY INTERDISCIPLINARY REVIEWS-COMPUTATIONAL STATISTICS, 2020, 12 (02)
  • [38] Hyperspectral Image Compression and Super-Resolution Using Tensor Decomposition Learning
    Aidini, A.
    Giannopoulos, M.
    Pentari, A.
    Fotiadou, K.
    Tsakalides, P.
    CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 1369 - 1373
  • [39] Tensor Denoising Using Low-Rank Tensor Train Decomposition
    Gong, Xiao
    Chen, Wei
    Chen, Jie
    Ai, Bo
    IEEE SIGNAL PROCESSING LETTERS, 2020, 27 : 1685 - 1689
  • [40] Rank minimization on tensor ring: an efficient approach for tensor decomposition and completion
    Yuan, Longhao
    Li, Chao
    Cao, Jianting
    Zhao, Qibin
    MACHINE LEARNING, 2020, 109 (03) : 603 - 622