On energy complexity of fully-connected layers

被引:0
|
作者
Sima, Jiri [1 ]
Cabessa, Jeremie [2 ]
Vidnerova, Petra [1 ]
机构
[1] Czech Acad Sci, Inst Comp Sci, Vodarenskou vezi 271-2, Prague 8, Czech Republic
[2] Univ Paris Saclay, Univ Versailles St Quentin UVSQ, DAVID Lab, 45 Ave Etats Unis, F-78035 Versailles, France
关键词
Deep neural networks; Convolutional neural networks; Fully-connected layer; Energy complexity; Energy consumption; Dataflow; NEURAL-NETWORKS;
D O I
10.1016/j.neunet.2024.106419
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The massive increase in the size of deep neural networks (DNNs) is accompanied by a significant increase in energy consumption of their hardware implementations which is critical for their widespread deployment in low -power mobile devices. In our previous work, an abstract hardware -independent model of energy complexity for convolutional neural networks (CNNs) has been proposed and experimentally validated. Based on this model, we provide a theoretical analysis of energy complexity related to the computation of a fullyconnected layer when its inputs, outputs, and weights are transferred between two kinds of memories (DRAM and Buffer). First, we establish a general lower bound on this energy complexity. Then, we present two dataflows and calculate their energy costs to achieve the corresponding upper bounds. In the case of a partitioned Buffer, we prove by the weak duality theorem from linear programming that the lower and upper bounds coincide up to an additive constant, and therefore establish the optimal energy complexity. Finally, the asymptotically optimal quadratic energy complexity of fully -connected layers is experimentally validated by estimating their energy consumption on the Simba and Eyeriss hardware.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Energy Complexity of Fully-Connected Layers
    Sima, Jiri
    Cabessa, Jeremie
    ADVANCES IN COMPUTATIONAL INTELLIGENCE, IWANN 2023, PT I, 2023, 14134 : 3 - 15
  • [2] Clique Here: On the Distributed Complexity in Fully-Connected Networks
    Applebaum, Benny
    Kowalski, Dariusz R.
    Patt-Shamir, Boaz
    Rosen, Adi
    PARALLEL PROCESSING LETTERS, 2016, 26 (01)
  • [3] Factorized Dynamic Fully-Connected Layers for Neural Networks
    Babiloni, Francesca
    Tanay, Thomas
    Deng, Jiankang
    Maggioni, Matteo
    Zafeiriou, Stefanos
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 1366 - 1375
  • [4] Overfitting remedy by sparsifying regularization on fully-connected layers of CNNs
    Xu, Qi
    Zhang, Ming
    Gu, Zonghua
    Pan, Gang
    NEUROCOMPUTING, 2019, 328 : 69 - 74
  • [5] The Unreasonable Effectiveness of Fully-Connected Layers for Low-Data Regimes
    Kocsis, Peter
    Sukenik, Peter
    Braso, Guillem
    Niessner, Matthias
    Leal-Taixe, Laura
    Elezi, Ismail
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [6] Structured Sparse Fully-Connected Layers in the CNNs and its GPU Acceleration
    Matsumura, Naoki
    Ito, Yasuaki
    Nakano, Koji
    Kasagi, Akihiko
    Tabaru, Tsuguchika
    2019 SEVENTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING WORKSHOPS (CANDARW 2019), 2019, : 148 - 154
  • [7] An Energy Efficient MAC Protocol for Fully-Connected Wireless Networks
    Malekshan, Kamal Rahimi
    Zhuang, Weihua
    2013 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2013, : 1538 - 1542
  • [8] Linear low energy excitations in fully-connected models of glasses
    Franz, Silvio
    Nicoletti, Flavio
    Ricci-Tersenghi, Federico
    JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2022, 2022 (05):
  • [9] Fully-connected networks with local connections
    P. E. Kornilovitch
    R. N. Bicknell
    J. S. Yeo
    Applied Physics A, 2009, 95 : 999 - 1004
  • [10] Fully-connected bond percolation on Zd
    Dereudre, David
    PROBABILITY THEORY AND RELATED FIELDS, 2022, 183 (1-2) : 547 - 579