Dense for the Price of Sparse: Improved Performance of Sparsely Initialized Networks via a Subspace Offset

被引:0
|
作者
Price, Ilan [1 ,2 ]
Tanner, Jared [1 ,2 ]
机构
[1] Univ Oxford, Math Inst, Oxford, England
[2] Alan Turing Inst, London, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
That neural networks may be pruned to high sparsities and retain high accuracy is well established. Recent research efforts focus on pruning immediately after initialization so as to allow the computational savings afforded by sparsity to extend to the training process. In this work, we introduce a new 'DCT plus Sparse' layer architecture, which maintains information propagation and trainability even with as little as 0.01% trainable kernel parameters remaining We show that standard training of networks built with these layers, and pruned at initialization, achieves state-of-the-art accuracy for extreme sparsities on a variety of benchmark network architectures and datasets. Moreover, these results are achieved using only simple heuristics to determine the locations of the trainable parameters in the network, and thus without having to initially store or compute with the full, unpruned network, as is required by competing prune-at-initialization algorithms. Switching from standard sparse layers to DCT plus Sparse layers does not increase the storage footprint of a network and incurs only a small additional computational overhead.
引用
收藏
页数:10
相关论文
共 25 条
  • [1] Sparse and dense hybrid representation via subspace modeling for dynamic MRI
    Liu, Qiegen
    Wang, Shanshan
    Liang, Dong
    [J]. COMPUTERIZED MEDICAL IMAGING AND GRAPHICS, 2017, 56 : 24 - 37
  • [2] An Improved RIP-Based Performance Guarantee for Sparse Signal Reconstruction via Subspace Pursuit
    Chang, Ling-Hua
    Wu, Jwo-Yuh
    [J]. 2014 IEEE 8TH SENSOR ARRAY AND MULTICHANNEL SIGNAL PROCESSING WORKSHOP (SAM), 2014, : 405 - 408
  • [3] Subspace clustering via stacked independent subspace analysis networks with sparse prior information
    Wu, Zongze
    Su, Chunchen
    Yin, Ming
    Ren, Zhigang
    Xie, Shengli
    [J]. PATTERN RECOGNITION LETTERS, 2021, 146 : 165 - 171
  • [4] An improved stopping condition guarantee recovery of sparse signal via Subspace Pursuit method
    Tawfic, Israa Sh.
    Kayhan, Sema
    [J]. ISA TRANSACTIONS, 2017, 70 : 149 - 160
  • [5] Improved sparse representation algorithm for face recognition via dense SIFT feature alignment
    Zhou, Quan
    Wei, Xin
    Chen, Jian-Xin
    Zheng, Bao-Yu
    [J]. Dianzi Yu Xinxi Xuebao/Journal of Electronics and Information Technology, 2015, 37 (08): : 1913 - 1919
  • [6] Training of sparse and dense deep neural networks: Fewer parameters, same performance
    Chicchi, Lorenzo
    Giambagli, Lorenzo
    Buffoni, Lorenzo
    Carletti, Timoteo
    Ciavarella, Marco
    Fanelli, Duccio
    [J]. PHYSICAL REVIEW E, 2021, 104 (05)
  • [7] Towards understanding residual and dilated dense neural networks via convolutional sparse coding
    Zhiyang Zhang
    Shihua Zhang
    [J]. National Science Review, 2021, 8 (03) : 127 - 139
  • [8] Towards understanding residual and dilated dense neural networks via convolutional sparse coding
    Zhang, Zhiyang
    Zhang, Shihua
    [J]. NATIONAL SCIENCE REVIEW, 2021, 8 (03)
  • [9] Improved Subspace K-Means Performance via a Randomized Matrix Decomposition
    Vannoy, Trevor
    Senecal, Jacob
    Strnadova-Neeley, Veronika
    [J]. 2019 7TH IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (IEEE GLOBALSIP), 2019,
  • [10] SPARSE-VIEW MEDICAL TOMOSYNTHESIS VIA MIXED SCALE DENSE CONVOLUTIONAL FRAMELET NETWORKS
    Alves Pereira, Luis F.
    De Beenhouwer, Jan
    Sijbers, Jan
    [J]. 2023 IEEE 20TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING, ISBI, 2023,