Efficient architecture for deep neural networks with heterogeneous sensitivity

被引:2
|
作者
Cho, Hyunjoong [1 ]
Jang, Jinhyeok [2 ]
Lee, Chanhyeok [1 ]
Yang, Seungjoon [1 ]
机构
[1] Ulsan Natl Inst Sci & Technol UNIST, Sch Elect & Comp Engn, Ulsan, South Korea
[2] Elect & Telecommun Res Inst ETRI, Daejeon, South Korea
基金
新加坡国家研究基金会;
关键词
Deep neural networks; Efficient architecture; Heterogeneous sensitivity; Constrained optimization; Simultaneous regularization parameter selection; L-CURVE; REGULARIZATION;
D O I
10.1016/j.neunet.2020.10.017
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this study, we present a neural network that consists of nodes with heterogeneous sensitivity. Each node in a network is assigned a variable that determines the sensitivity with which it learns to perform a given task. The network is trained via a constrained optimization that maximizes the sparsity of the sensitivity variables while ensuring optimal network performance. As a result, the network learns to perform a given task using only a few sensitive nodes. Insensitive nodes, which are nodes with zero sensitivity, can be removed from a trained network to obtain a computationally efficient network. Removing zero-sensitivity nodes has no effect on the performance of the network because the network has already been trained to perform the task without them. The regularization parameter used to solve the optimization problem was simultaneously found during the training of the networks. To validate our approach, we designed networks with computationally efficient architectures for various tasks such as autoregression, object recognition, facial expression recognition, and object detection using various datasets. In our experiments, the networks designed by our proposed method provided the same or higher performances but with far less computational complexity. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页码:95 / 106
页数:12
相关论文
共 50 条
  • [21] LiteLSTM Architecture for Deep Recurrent Neural Networks
    Elsayed, Nelly
    ElSayed, Zag
    Maida, Anthony S.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1304 - 1308
  • [22] An Architecture to Accelerate Convolution in Deep Neural Networks
    Ardakani, Arash
    Condo, Carlo
    Ahmadi, Mehdi
    Gross, Warren J.
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2018, 65 (04) : 1349 - 1362
  • [23] Hardware Architecture Exploration for Deep Neural Networks
    Wenqi Zheng
    Yangyi Zhao
    Yunfan Chen
    Jinhong Park
    Hyunchul Shin
    Arabian Journal for Science and Engineering, 2021, 46 : 9703 - 9712
  • [24] Deploying and Optimizing Convolutional Neural Networks on Heterogeneous Architecture
    Jiang, Junning
    Cai, Liang
    Dong, Feng
    Yu, Kehua
    Chen, Ke
    Qu, Wei
    Jiang, Jianfei
    2019 IEEE 13TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2019,
  • [25] Fusion based Heterogeneous Convolutional Neural Networks Architecture
    Komish, David
    Ezekiel, Soundararajan
    Comacchia, Maria
    2018 IEEE APPLIED IMAGERY PATTERN RECOGNITION WORKSHOP (AIPR), 2018,
  • [26] An efficient synaptic architecture for artificial neural networks
    Boybat, Irem
    Le Gallo, Manuel
    Nandakumar, S. R.
    Moraitis, Timoleon
    Tuma, Tomas
    Rajendran, Bipin
    Leblebici, Yusuf
    Sebastian, Abu
    Eleftheriou, Evangelos
    2017 17TH NON-VOLATILE MEMORY TECHNOLOGY SYMPOSIUM (NVMTS), 2017,
  • [27] RESPARC: A Reconfigurable and Energy-Efficient Architecture with Memristive Crossbars for Deep Spiking Neural Networks
    Ankit, Aayush
    Sengupta, Abhronil
    Panda, Priyadarshini
    Roy, Kaushik
    PROCEEDINGS OF THE 2017 54TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2017,
  • [28] An Efficient Accelerator for Deep Convolutional Neural Networks
    Kuo, Yi-Xian
    Lai, Yeong-Kang
    2020 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TAIWAN), 2020,
  • [29] Efficient Deep Neural Networks for Edge Computing
    Alnemari, Mohammed
    Bagherzadeh, Nader
    2019 IEEE INTERNATIONAL CONFERENCE ON EDGE COMPUTING (IEEE EDGE), 2019, : 1 - 7
  • [30] Efficient Model Averaging for Deep Neural Networks
    Opitz, Michael
    Possegger, Horst
    Bischof, Horst
    COMPUTER VISION - ACCV 2016, PT II, 2017, 10112 : 205 - 220