Sensitivity Analysis for Deep Learning: Ranking Hyper-parameter Influence

被引:20
|
作者
Taylor, Rhian [1 ]
Ojha, Varun [1 ]
Martino, Ivan [2 ]
Nicosia, Giuseppe [3 ]
机构
[1] Univ Reading, Dept Comp Sci, Reading, Berks, England
[2] KTH Royal Inst Technol, Stockholm, Sweden
[3] Univ Cambridge, Cambridge, England
关键词
Sensitivity Analysis; Deep Learning; Hyper-parameter Tuning; Hyper-parameter rank; Hyper-parameter Influence; NEURAL-NETWORKS; UNCERTAINTY;
D O I
10.1109/ICTAI52525.2021.00083
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a novel approach to rank Deep Learning (DL) hyper-parameters through the application of Sensitivity Analysis (SA). DL hyper-parameter tuning is crucial to model accuracy however, choosing optimal values for each parameter is time and resource-intensive. SA provides a quantitative measure by which hyper-parameters can be ranked in terms of contribution to model accuracy. Learning rate decay was ranked highest, with model performance being sensitive to this parameter regardless of architecture or dataset. The influence of a model's initial learning rate was proven to be low, contrary to the literature. Additionally, the importance of a parameter is closely linked to model architecture. Shallower models showed susceptibility to hyper-parameters affecting the stochasticity of the learning process whereas deeper models showed sensitivity to hyper-parameters affecting the convergence speed. Furthermore, the complexity of the dataset can affect the margin of separation between the sensitivity measures of the most and the least influential parameters, making the most influential hyper-parameter an ideal candidate for tuning compared to the other parameters.
引用
收藏
页码:512 / 516
页数:5
相关论文
共 50 条
  • [41] A review of automatic selection methods for machine learning algorithms and hyper-parameter values
    Luo, Gang
    NETWORK MODELING AND ANALYSIS IN HEALTH INFORMATICS AND BIOINFORMATICS, 2016, 5 (01):
  • [42] Gradient Hyper-parameter Optimization for Manifold Regularization
    Becker, Cassiano O.
    Ferreira, Paulo A. V.
    2013 12TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2013), VOL 2, 2013, : 339 - 344
  • [43] Hyper-parameter Tuning for Progressive Learning and its Application to Network Cyber Security
    Karn, Rupesh Raj
    Ziegler, Matthew
    Jung, Jinwook
    Elfadel, Ibrahim M.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1220 - 1224
  • [44] Total Variation with Automatic Hyper-Parameter Estimation
    Nascimento, Jacinto
    Sanches, Joao
    2008 30TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY, VOLS 1-8, 2008, : 443 - +
  • [45] A study on depth classification of defects by machine learning based on hyper-parameter search
    Chen, Haoze
    Zhang, Zhijie
    Yin, Wuliang
    Zhao, Chenyang
    Wang, Fengxiang
    Li, Yanfeng
    MEASUREMENT, 2022, 189
  • [46] Hyper-Parameter in Hidden Markov Random Field
    Lim, Johan
    Yu, Donghyeon
    Pyun, Kyungsuk
    KOREAN JOURNAL OF APPLIED STATISTICS, 2011, 24 (01) : 177 - 183
  • [47] Bayesian Optimization for Accelerating Hyper-parameter Tuning
    Vu Nguyen
    2019 IEEE SECOND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND KNOWLEDGE ENGINEERING (AIKE), 2019, : 302 - 305
  • [48] Exploring Parameter and Hyper-Parameter Spaces of Neuroscience Models on High Performance Computers With Learning to Learn
    Yegenoglu, Alper
    Subramoney, Anand
    Hater, Thorsten
    Jimenez-Romero, Cristian
    Klijn, Wouter
    Martin, AaronPerez
    van der Vlag, Michiel
    Herty, Michael
    Morrison, Abigail
    Diaz-Pier, Sandra
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2022, 16
  • [49] On hyper-parameter selection for guaranteed convergence of RMSProp
    Liu, Jinlan
    Xu, Dongpo
    Zhang, Huisheng
    Mandic, Danilo
    COGNITIVE NEURODYNAMICS, 2022, 18 (6) : 3227 - 3237
  • [50] A Comparative study of Hyper-Parameter Optimization Tools
    Shekhar, Shashank
    Bansode, Adesh
    Salim, Asif
    2021 IEEE ASIA-PACIFIC CONFERENCE ON COMPUTER SCIENCE AND DATA ENGINEERING (CSDE), 2021,