Efficient architecture for deep neural networks with heterogeneous sensitivity

被引:2
|
作者
Cho, Hyunjoong [1 ]
Jang, Jinhyeok [2 ]
Lee, Chanhyeok [1 ]
Yang, Seungjoon [1 ]
机构
[1] Ulsan Natl Inst Sci & Technol UNIST, Sch Elect & Comp Engn, Ulsan, South Korea
[2] Elect & Telecommun Res Inst ETRI, Daejeon, South Korea
基金
新加坡国家研究基金会;
关键词
Deep neural networks; Efficient architecture; Heterogeneous sensitivity; Constrained optimization; Simultaneous regularization parameter selection; L-CURVE; REGULARIZATION;
D O I
10.1016/j.neunet.2020.10.017
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this study, we present a neural network that consists of nodes with heterogeneous sensitivity. Each node in a network is assigned a variable that determines the sensitivity with which it learns to perform a given task. The network is trained via a constrained optimization that maximizes the sparsity of the sensitivity variables while ensuring optimal network performance. As a result, the network learns to perform a given task using only a few sensitive nodes. Insensitive nodes, which are nodes with zero sensitivity, can be removed from a trained network to obtain a computationally efficient network. Removing zero-sensitivity nodes has no effect on the performance of the network because the network has already been trained to perform the task without them. The regularization parameter used to solve the optimization problem was simultaneously found during the training of the networks. To validate our approach, we designed networks with computationally efficient architectures for various tasks such as autoregression, object recognition, facial expression recognition, and object detection using various datasets. In our experiments, the networks designed by our proposed method provided the same or higher performances but with far less computational complexity. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页码:95 / 106
页数:12
相关论文
共 50 条
  • [41] COSY: An Energy-Efficient Hardware Architecture for Deep Convolutional Neural Networks based on Systolic Array
    Yin, Chen
    Chen, Qiang
    Tian, Miren
    Ji, Mohan
    Zou, Chenglong
    Wang, Yin'an
    Wang, Bo
    2017 IEEE 23RD INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2017, : 180 - 189
  • [42] Energy-Efficient Architecture for FPGA-based Deep Convolutional Neural Networks with Binary Weights
    Duan, Yunzhi
    Li, Shuai
    Zhang, Ruipeng
    Wang, Qi
    Chen, Jienan
    Sobelman, Gerald E.
    2018 IEEE 23RD INTERNATIONAL CONFERENCE ON DIGITAL SIGNAL PROCESSING (DSP), 2018,
  • [43] Optimization of neural networks architecture for impact sensitivity of energetic molecules
    Cho, SG
    No, KT
    Goh, EM
    Kim, JK
    Shin, JH
    Joo, YD
    Seong, S
    BULLETIN OF THE KOREAN CHEMICAL SOCIETY, 2005, 26 (03): : 399 - 408
  • [44] Heterogeneous gradient computing optimization for scalable deep neural networks
    Sergio Moreno-Álvarez
    Mercedes E. Paoletti
    Juan A. Rico-Gallego
    Juan M. Haut
    The Journal of Supercomputing, 2022, 78 : 13455 - 13469
  • [45] DeepTLF: robust deep neural networks for heterogeneous tabular data
    Borisov, Vadim
    Broelemann, Klaus
    Kasneci, Enkelejda
    Kasneci, Gjergji
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2023, 16 (01) : 85 - 100
  • [46] Heterogeneous gradient computing optimization for scalable deep neural networks
    Moreno-Alvarez, Sergio
    Paoletti, Mercedes E.
    Rico-Gallego, Juan A.
    Haut, Juan M.
    JOURNAL OF SUPERCOMPUTING, 2022, 78 (11): : 13455 - 13469
  • [47] Deep heterogeneous network embedding based on Siamese Neural Networks
    Zhang, Chen
    Tang, Zhouhua
    Yu, Bin
    Xie, Yu
    Pan, Ke
    NEUROCOMPUTING, 2020, 388 : 1 - 11
  • [48] DeepTLF: robust deep neural networks for heterogeneous tabular data
    Vadim Borisov
    Klaus Broelemann
    Enkelejda Kasneci
    Gjergji Kasneci
    International Journal of Data Science and Analytics, 2023, 16 : 85 - 100
  • [49] Evolutionary Training of Deep Neural Networks on Heterogeneous Computing Environments
    Kalia, Subodh
    Mohan, Chilukuri K.
    Nemani, Ramakrishna
    PROCEEDINGS OF THE 2022 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2022, 2022, : 2318 - 2321
  • [50] Efficient Hardware Architecture of Softmax Layer in Deep Neural Network
    Hu, Ruofei
    Tian, Binren
    Yin, Shouyi
    Wei, Shaojun
    2018 IEEE 23RD INTERNATIONAL CONFERENCE ON DIGITAL SIGNAL PROCESSING (DSP), 2018,