Analysis of Layer Efficiency and Layer Reduction on Pre-trained Deep Learning Models

被引:0
|
作者
Nugraha, Brilian Tafjira [1 ]
Su, Shun-Feng [1 ]
机构
[1] NTUST, Taipei 106, Taiwan
关键词
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Recent technologies in the deep learning area enable many industries and practitioners fastening the development processes of their products. However, deep learning still encounters some potential issues like overfitting and huge size. The huge size greatly constrains performance and portability of the deep learning model in embedded devices with limited environments. Due to the paradigm of it mixed with the meaning of "deep" layers, many researchers tend to derive the pre-trained model into building deeper layers to solve their problems without knowing whether they are actually needed or not. To address these issues, we exploit the activation and gradient output and weight in each layer of the pre-trained models to measure its efficiencies. By exploiting them, we estimate the efficiencies using our measurements and compare it with the manual layer reduction to validate the most relevant method. We also use the method for continuous layer reductions for validation. With this approach, we save up to 12x and 26x of the time of one manual layer reduction and re-training on VGG-16 and custom AlexNet respectively.
引用
收藏
页数:6
相关论文
共 50 条
  • [11] Mass detection in mammograms using pre-trained deep learning models
    Agarwal, Richa
    Diaz, Oliver
    Llado, Xavier
    Marti, Robert
    14TH INTERNATIONAL WORKSHOP ON BREAST IMAGING (IWBI 2018), 2018, 10718
  • [12] An Approach to Run Pre-Trained Deep Learning Models on Grayscale Images
    Ahmad, Ijaz
    Shin, Seokjoo
    3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION (IEEE ICAIIC 2021), 2021, : 177 - 180
  • [13] Deep Compression of Pre-trained Transformer Models
    Wang, Naigang
    Liu, Chi-Chun
    Venkataramani, Swagath
    Sen, Sanchari
    Chen, Chia-Yu
    El Maghraoui, Kaoutar
    Srinivasan, Vijayalakshmi
    Chang, Leland
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [14] Classification Efficiency of Pre-Trained Deep CNN Models on Camera Trap Images
    Stancic, Adam
    Vyroubal, Vedran
    Slijepcevic, Vedran
    JOURNAL OF IMAGING, 2022, 8 (02)
  • [15] Learning to Modulate pre-trained Models in RL
    Schmied, Thomas
    Hofmarcher, Markus
    Paischer, Fabian
    Pascanu, Razvan
    Hochreiter, Sepp
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [16] Benchmark analysis of various pre-trained deep learning models on ASSIRA cats and dogs dataset
    Galib Muhammad Shahriar Himel
    Md. Masudul Islam
    Journal of Umm Al-Qura University for Engineering and Architecture, 2025, 16 (1): : 134 - 149
  • [17] Continual Learning with Pre-Trained Models: A Survey
    Zhou, Da-Wei
    Sun, Hai-Long
    Ning, Jingyi
    Ye, Han-Jia
    Zhan, De-Chuan
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 8363 - 8371
  • [18] An Ensemble Voting Method of Pre-Trained Deep Learning Models for Orchid Recognition
    Ou, Chia-Ho
    Hu, Yi-Nuo
    Jiang, Dong-Jie
    Liao, Po-Yen
    2023 IEEE INTERNATIONAL SYSTEMS CONFERENCE, SYSCON, 2023,
  • [19] Kurdish Sign Language Recognition Using Pre-Trained Deep Learning Models
    Alsaud, Ali A.
    Yousif, Raghad Z.
    Aziz, Marwan. M.
    Kareem, Shahab W.
    Maho, Amer J.
    JOURNAL OF ELECTRICAL SYSTEMS, 2024, 20 (06) : 1334 - 1344
  • [20] Integration of pre-trained protein language models into geometric deep learning networks
    Wu, Fang
    Wu, Lirong
    Radev, Dragomir
    Xu, Jinbo
    Li, Stan Z.
    COMMUNICATIONS BIOLOGY, 2023, 6 (01)