Analysis of Layer Efficiency and Layer Reduction on Pre-trained Deep Learning Models

被引:0
|
作者
Nugraha, Brilian Tafjira [1 ]
Su, Shun-Feng [1 ]
机构
[1] NTUST, Taipei 106, Taiwan
关键词
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Recent technologies in the deep learning area enable many industries and practitioners fastening the development processes of their products. However, deep learning still encounters some potential issues like overfitting and huge size. The huge size greatly constrains performance and portability of the deep learning model in embedded devices with limited environments. Due to the paradigm of it mixed with the meaning of "deep" layers, many researchers tend to derive the pre-trained model into building deeper layers to solve their problems without knowing whether they are actually needed or not. To address these issues, we exploit the activation and gradient output and weight in each layer of the pre-trained models to measure its efficiencies. By exploiting them, we estimate the efficiencies using our measurements and compare it with the manual layer reduction to validate the most relevant method. We also use the method for continuous layer reductions for validation. With this approach, we save up to 12x and 26x of the time of one manual layer reduction and re-training on VGG-16 and custom AlexNet respectively.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] Comparative Analysis of ImageNet Pre-Trained Deep Learning Models and DINOv2 in Medical Imaging Classification
    Huang, Yuning
    Zou, Jingchen
    Meng, Lanxi
    Yue, Xin
    Zhao, Qing
    Li, Jianqiang
    Song, Changwei
    Jimenez, Gabriel
    Li, Shaowu
    Fu, Guanghui
    2024 IEEE 48TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC 2024, 2024, : 297 - 305
  • [42] On Pre-trained Image Features and Synthetic Images for Deep Learning
    Hinterstoisser, Stefan
    Lepetit, Vincent
    Wohlhart, Paul
    Konolige, Kurt
    COMPUTER VISION - ECCV 2018 WORKSHOPS, PT I, 2019, 11129 : 682 - 697
  • [43] Federated Learning from Pre-Trained Models: A Contrastive Learning Approach
    Tan, Yue
    Long, Guodong
    Ma, Jie
    Liu, Lu
    Zhou, Tianyi
    Jiang, Jing
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [44] RanPAC: Random Projections and Pre-trained Models for Continual Learning
    McDonnell, Mark D.
    Gong, Dong
    Parveneh, Amin
    Abbasnejad, Ehsan
    van den Hengel, Anton
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [45] Context Analysis for Pre-trained Masked Language Models
    Lai, Yi-An
    Lalwani, Garima
    Zhang, Yi
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 3789 - 3804
  • [46] CODEEDITOR: Learning to Edit Source Code with Pre-trained Models
    Li, Jia
    Li, Ge
    Li, Zhuo
    Jin, Zhi
    Hu, Xing
    Zhang, Kechi
    Fu, Zhiyi
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2023, 32 (06)
  • [47] Zero-shot Learning for Subdiscrimination in Pre-trained Models
    Dominguez-Mateos, Francisco
    O'Brien, Vincent
    Garland, James
    Furlong, Ryan
    Palacios-Alonso, Daniel
    JOURNAL OF UNIVERSAL COMPUTER SCIENCE, 2025, 31 (01) : 93 - 110
  • [48] EVALUATION OF DIFFERENT PARAMETERS FOR PLANT CLASSIFICATION BY PRE-TRAINED DEEP LEARNING MODELS WITH BIGEARTHNET DATASET
    Naali, F.
    Alipour-Fard, T.
    Arefi, H.
    ISPRS GEOSPATIAL CONFERENCE 2022, JOINT 6TH SENSORS AND MODELS IN PHOTOGRAMMETRY AND REMOTE SENSING, SMPR/4TH GEOSPATIAL INFORMATION RESEARCH, GIRESEARCH CONFERENCES, VOL. 10-4, 2023, : 569 - 574
  • [49] Deep learning approaches for online signature authentication: a comparative study of pre-trained CNN models
    Swamy, M. Ranga
    Vijayalakshmi, P.
    Rajendran, V
    ENGINEERING RESEARCH EXPRESS, 2025, 7 (01):
  • [50] Collaborative Learning across Heterogeneous Systems with Pre-Trained Models
    Hoang, Trong Nghia
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 20, 2024, : 22668 - 22668