The Loss Surface of Deep and Wide Neural Networks

被引:0
|
作者
Quynh Nguyen [1 ]
Hein, Matthias [1 ]
机构
[1] Saarland Univ, Dept Math & Comp Sci, Saarbrucken, Germany
基金
欧洲研究理事会;
关键词
LOCAL MINIMA;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While the optimization problem behind deep neural networks is highly non-convex, it is frequently observed in practice that training deep networks seems possible without getting stuck in suboptimal points. It has been argued that this is the case as all local minima are close to being globally optimal. We show that this is (almost) true, in fact almost all local minima are globally optimal, for a fully connected network with squared loss and analytic activation function given that the number of hidden units of one layer of the network is larger than the number of training points and the network structure from this layer on is pyramidal.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Wide and Deep Neural Networks for Automatic Liver Segmentation in Liver Iron Quantification
    Liu, M.
    Roytlender, A.
    Jambawalikar, S.
    MEDICAL PHYSICS, 2019, 46 (06) : E161 - E161
  • [42] Non-attracting regions of local minima in deep and wide neural networks
    Petzka, Henning
    Sminchisescu, Cristian
    Journal of Machine Learning Research, 2021, 22 : 1 - 34
  • [43] Deep Versus Wide Convolutional Neural Networks for Object Recognition on Neuromorphic System
    Alom, Md Zahangir
    Josue, Theodore
    Rahman, Md Nayim
    Mitchell, Will
    Yakopcic, Chris
    Taha, Tarek M.
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [44] Wide Neural Networks as Gaussian Processes: Lessons from Deep Equilibrium Models
    Gao, Tianxiang
    Huo, Xiaokai
    Liu, Hailiang
    Gao, Hongyang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [45] Non-attracting Regions of Local Minima in Deep and Wide Neural Networks
    Petzka, Henning
    Sminchisescu, Cristian
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [46] On Generalization Bounds for Deep Networks Based on Loss Surface Implicit Regularization
    Imaizumi, Masaaki
    Schmidt-Hieber, Johannes
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2023, 69 (02) : 1203 - 1223
  • [47] Making Deep Neural Networks Robust to Label Noise: a Loss Correction Approach
    Patrini, Giorgio
    Rozza, Alessandro
    Menon, Aditya Krishna
    Nock, Richard
    Qu, Lizhen
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2233 - 2241
  • [48] The Loss Landscape of Deep Linear Neural Networks: a Second-order Analysis
    Achour, El Mehdi
    Malgouyres, Francois
    Gerchinovitz, Sebastien
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 76
  • [49] Path Loss Prediction Based on Deep Neural Networks Considering Fresnel Zone
    Wang, Shirui
    He, Ruisi
    Wu, Tong
    Yang, Jingya
    Yang, Mi
    Zhou, Shun
    Wang, Chenlong
    Ai, Bo
    2024 INTERNATIONAL CONFERENCE ON UBIQUITOUS COMMUNICATION, UCOM 2024, 2024, : 11 - 15
  • [50] Improving the Robustness of Deep Neural Networks via Adversarial Training with Triplet Loss
    Li, Pengcheng
    Yi, Jinfeng
    Zhou, Bowen
    Zhang, Lijun
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2909 - 2915