Analysis of Layer Efficiency and Layer Reduction on Pre-trained Deep Learning Models

被引:0
|
作者
Nugraha, Brilian Tafjira [1 ]
Su, Shun-Feng [1 ]
机构
[1] NTUST, Taipei 106, Taiwan
关键词
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Recent technologies in the deep learning area enable many industries and practitioners fastening the development processes of their products. However, deep learning still encounters some potential issues like overfitting and huge size. The huge size greatly constrains performance and portability of the deep learning model in embedded devices with limited environments. Due to the paradigm of it mixed with the meaning of "deep" layers, many researchers tend to derive the pre-trained model into building deeper layers to solve their problems without knowing whether they are actually needed or not. To address these issues, we exploit the activation and gradient output and weight in each layer of the pre-trained models to measure its efficiencies. By exploiting them, we estimate the efficiencies using our measurements and compare it with the manual layer reduction to validate the most relevant method. We also use the method for continuous layer reductions for validation. With this approach, we save up to 12x and 26x of the time of one manual layer reduction and re-training on VGG-16 and custom AlexNet respectively.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] A Performance Comparison of Pre-trained Deep Learning Models to Classify Brain Tumor
    Diker, Aykut
    IEEE EUROCON 2021 - 19TH INTERNATIONAL CONFERENCE ON SMART TECHNOLOGIES, 2021, : 246 - 249
  • [22] Integration of pre-trained protein language models into geometric deep learning networks
    Fang Wu
    Lirong Wu
    Dragomir Radev
    Jinbo Xu
    Stan Z. Li
    Communications Biology, 6
  • [23] Enhancement of Pre-Trained Deep Learning Models to Improve Brain Tumor Classification
    Ullah Z.
    Odeh A.
    Khattak I.
    Hasan M.A.
    Informatica (Slovenia), 2023, 47 (06): : 165 - 172
  • [24] HadSkip: Homotopic and Adaptive Layer Skipping of Pre-trained Language Models for Efficient Inference
    Wang, Haoyu
    Wang, Yaqing
    Liu, Tianci
    Zhao, Tuo
    Gao, Jing
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 4283 - 4294
  • [25] Selective Layer Tuning and Performance Study of Pre-Trained Models Using Genetic Algorithm
    Jeong, Jae-Cheol
    Yu, Gwang-Hyun
    Song, Min-Gyu
    Dang Thanh Vu
    Le Hoang Anh
    Jung, Young-Ae
    Choi, Yoon-A
    Um, Tai-Won
    Kim, Jin-Young
    ELECTRONICS, 2022, 11 (19)
  • [26] Deep Entity Matching with Pre-Trained Language Models
    Li, Yuliang
    Li, Jinfeng
    Suhara, Yoshihiko
    Doan, AnHai
    Tan, Wang-Chiew
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2020, 14 (01): : 50 - 60
  • [27] Integrating Pre-Trained Language Model With Physical Layer Communications
    Lee, Ju-Hyung
    Lee, Dong-Ho
    Lee, Joohan
    Pujara, Jay
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2024, 23 (11) : 17266 - 17278
  • [28] Towards Inadequately Pre-trained Models in Transfer Learning
    Deng, Andong
    Li, Xingjian
    Hu, Di
    Wang, Tianyang
    Xiong, Haoyi
    Xu, Cheng-Zhong
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 19340 - 19351
  • [29] Transfer learning with pre-trained conditional generative models
    Yamaguchi, Shin'ya
    Kanai, Sekitoshi
    Kumagai, Atsutoshi
    Chijiwa, Daiki
    Kashima, Hisashi
    MACHINE LEARNING, 2025, 114 (04)
  • [30] The severity level classification of Fusarium wilt of chickpea by pre-trained deep learning models
    Tolga Hayit
    Ali Endes
    Fatma Hayit
    Journal of Plant Pathology, 2024, 106 (1) : 93 - 105