Analysis of Layer Efficiency and Layer Reduction on Pre-trained Deep Learning Models

被引:0
|
作者
Nugraha, Brilian Tafjira [1 ]
Su, Shun-Feng [1 ]
机构
[1] NTUST, Taipei 106, Taiwan
关键词
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Recent technologies in the deep learning area enable many industries and practitioners fastening the development processes of their products. However, deep learning still encounters some potential issues like overfitting and huge size. The huge size greatly constrains performance and portability of the deep learning model in embedded devices with limited environments. Due to the paradigm of it mixed with the meaning of "deep" layers, many researchers tend to derive the pre-trained model into building deeper layers to solve their problems without knowing whether they are actually needed or not. To address these issues, we exploit the activation and gradient output and weight in each layer of the pre-trained models to measure its efficiencies. By exploiting them, we estimate the efficiencies using our measurements and compare it with the manual layer reduction to validate the most relevant method. We also use the method for continuous layer reductions for validation. With this approach, we save up to 12x and 26x of the time of one manual layer reduction and re-training on VGG-16 and custom AlexNet respectively.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] DenseNet-201 and Xception Pre-Trained Deep Learning Models for Fruit Recognition
    Salim, Farsana
    Saeed, Faisal
    Basurra, Shadi
    Qasem, Sultan Noman
    Al-Hadhrami, Tawfik
    ELECTRONICS, 2023, 12 (14)
  • [32] Active Learning for Sequence Tagging with Deep Pre-trained Models and Bayesian Uncertainty Estimates
    Shelmanov, Artem
    Puzyrev, Dmitri
    Kupriyanova, Lyubov
    Belyakov, Denis
    Larionov, Daniil
    Khromov, Nikita
    Kozlova, Olga
    Artemova, Ekaterina
    Dylov, Dmitry, V
    Panchenko, Alexander
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1698 - 1712
  • [33] Active Learning with Deep Pre-trained Models for Sequence Tagging of Clinical and Biomedical Texts
    Shelmanov, Artem
    Liventsev, Vadini
    Kireev, Danil
    Khromov, Nikita
    Panchenko, Alexander
    Fedulova, Irina
    Dylov, Dmitry, V
    2019 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE (BIBM), 2019, : 482 - 489
  • [34] The use of pre-trained deep learning models for the photographic assessment of donor livers for transplantation
    Ugail, Hassan
    Abubakar, Aliyu
    Elmahmudi, Ali
    Wilson, Colin
    Thomson, Brian
    ARTIFICIAL INTELLIGENCE SURGERY, 2022, 2 (02): : 101 - 119
  • [35] Person Re-Identification with Pre-trained Deep Learning Models and Attribute Correlations
    Ngoc-Bao Nguyen
    Vu-Hoang Nguyen
    Tien Do
    Thanh Due Ngo
    2016 INTERNATIONAL SYMPOSIUM ON ELECTRONICS AND SMART DEVICES (ISESD), 2016, : 242 - 247
  • [36] Improved White Blood Cells Classification Based on Pre-trained Deep Learning Models
    Mohamed, Ensaf H.
    El-Behaidy, Wessam H.
    Khoriba, Ghada
    Li, Jie
    JOURNAL OF COMMUNICATIONS SOFTWARE AND SYSTEMS, 2020, 16 (01) : 37 - 45
  • [37] LMPred: predicting antimicrobial peptides using pre-trained language models and deep learning
    Dee, William
    Gromiha, Michael
    BIOINFORMATICS ADVANCES, 2022, 2 (01):
  • [38] The severity level classification of Fusarium wilt of chickpea by pre-trained deep learning models
    Hayit, Tolga
    Endes, Ali
    Hayit, Fatma
    JOURNAL OF PLANT PATHOLOGY, 2024, 106 (01) : 93 - 105
  • [39] Deep Fusing Pre-trained Models into Neural Machine Translation
    Weng, Rongxiang
    Yu, Heng
    Luo, Weihua
    Zhang, Min
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11468 - 11476
  • [40] Uncertainty Estimation and Reduction of Pre-trained Models for Text Regression
    Wang, Yuxia
    Beck, Daniel
    Baldwin, Timothy
    Verspoor, Karin
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 680 - 696