Sensitivity Analysis for Deep Learning: Ranking Hyper-parameter Influence

被引:20
|
作者
Taylor, Rhian [1 ]
Ojha, Varun [1 ]
Martino, Ivan [2 ]
Nicosia, Giuseppe [3 ]
机构
[1] Univ Reading, Dept Comp Sci, Reading, Berks, England
[2] KTH Royal Inst Technol, Stockholm, Sweden
[3] Univ Cambridge, Cambridge, England
关键词
Sensitivity Analysis; Deep Learning; Hyper-parameter Tuning; Hyper-parameter rank; Hyper-parameter Influence; NEURAL-NETWORKS; UNCERTAINTY;
D O I
10.1109/ICTAI52525.2021.00083
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a novel approach to rank Deep Learning (DL) hyper-parameters through the application of Sensitivity Analysis (SA). DL hyper-parameter tuning is crucial to model accuracy however, choosing optimal values for each parameter is time and resource-intensive. SA provides a quantitative measure by which hyper-parameters can be ranked in terms of contribution to model accuracy. Learning rate decay was ranked highest, with model performance being sensitive to this parameter regardless of architecture or dataset. The influence of a model's initial learning rate was proven to be low, contrary to the literature. Additionally, the importance of a parameter is closely linked to model architecture. Shallower models showed susceptibility to hyper-parameters affecting the stochasticity of the learning process whereas deeper models showed sensitivity to hyper-parameters affecting the convergence speed. Furthermore, the complexity of the dataset can affect the margin of separation between the sensitivity measures of the most and the least influential parameters, making the most influential hyper-parameter an ideal candidate for tuning compared to the other parameters.
引用
收藏
页码:512 / 516
页数:5
相关论文
共 50 条
  • [31] A new hyper-parameter optimization method for machine learning in fault classification
    Ye, Xingchen
    Gao, Liang
    Li, Xinyu
    Wen, Long
    APPLIED INTELLIGENCE, 2023, 53 (11) : 14182 - 14200
  • [32] Learning sparse linear dynamic networks in a hyper-parameter free setting
    Venkitaraman, Arun
    Hjalmarsson, Hakan
    Wahlberg, Bo
    IFAC PAPERSONLINE, 2020, 53 (02): : 82 - 86
  • [33] Hyper-Parameter Tuning for Graph Kernels via Multiple Kernel Learning
    Massimo, Carlo M.
    Navarin, Nicolo
    Sperduti, Alessandro
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT II, 2016, 9948 : 214 - 223
  • [34] Hyper-parameter Optimization for Machine-Learning based Electromagnetic Side-Channel Analysis
    Mukhtar, Naila
    Kong, Yinan
    2018 26TH INTERNATIONAL CONFERENCE ON SYSTEMS ENGINEERING (ICSENG 2018), 2018,
  • [35] Deep neural network hyper-parameter tuning through twofold genetic approach
    Kumar, Puneet
    Batra, Shalini
    Raman, Balasubramanian
    SOFT COMPUTING, 2021, 25 (13) : 8747 - 8771
  • [36] Variational Reinforcement Learning for Hyper-Parameter Tuning of Adaptive Evolutionary Algorithm
    Zhang, Haotian
    Sun, Jianyong
    Wang, Yuhao
    Shi, Jialong
    Xu, Zongben
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2023, 7 (05): : 1511 - 1526
  • [37] A new hyper-parameter optimization method for machine learning in fault classification
    Xingchen Ye
    Liang Gao
    Xinyu Li
    Long Wen
    Applied Intelligence, 2023, 53 : 14182 - 14200
  • [38] Deep neural network hyper-parameter tuning through twofold genetic approach
    Puneet Kumar
    Shalini Batra
    Balasubramanian Raman
    Soft Computing, 2021, 25 : 8747 - 8771
  • [39] Efficient Federated Learning with Adaptive Client-Side Hyper-Parameter Optimization
    Kundroo, Majid
    Kim, Taehong
    2023 IEEE 43RD INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS, ICDCS, 2023, : 973 - 974
  • [40] Hyper-Parameter Optimization Using MARS Surrogate for Machine-Learning Algorithms
    Li, Yangyang
    Liu, Guangyuan
    Lu, Gao
    Jiao, Licheng
    Marturi, Naresh
    Shang, Ronghua
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2020, 4 (03): : 287 - 297