The Loss Surface of Deep and Wide Neural Networks

被引:0
|
作者
Quynh Nguyen [1 ]
Hein, Matthias [1 ]
机构
[1] Saarland Univ, Dept Math & Comp Sci, Saarbrucken, Germany
基金
欧洲研究理事会;
关键词
LOCAL MINIMA;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While the optimization problem behind deep neural networks is highly non-convex, it is frequently observed in practice that training deep networks seems possible without getting stuck in suboptimal points. It has been argued that this is the case as all local minima are close to being globally optimal. We show that this is (almost) true, in fact almost all local minima are globally optimal, for a fully connected network with squared loss and analytic activation function given that the number of hidden units of one layer of the network is larger than the number of training points and the network structure from this layer on is pyramidal.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Dimensionality-Induced Information Loss of Outliers in Deep Neural Networks
    Uematsu, Kazuki
    Haruki, Kosuke
    Suzuki, Taiji
    Kimura, Mitsuhiro
    Takimoto, Takahiro
    Nakagawa, Hideyuki
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT I, ECML PKDD 2024, 2024, 14941 : 144 - 160
  • [32] Embedding Principle in Depth for the Loss Landscape Analysis of Deep Neural Networks
    Bai, Zhiwei
    Luo, Tao
    Xu, Zhi-Qin John
    Zhang, Yaoyu
    CSIAM TRANSACTIONS ON APPLIED MATHEMATICS, 2024, 5 (02): : 350 - 389
  • [33] Deeper Weight Pruning without Accuracy Loss in Deep Neural Networks
    Ahn, Byungmin
    Kim, Taewhan
    PROCEEDINGS OF THE 2020 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2020), 2020, : 73 - 78
  • [34] Empirical study on tangent loss function for classification with deep neural networks
    Zhang, Xu
    Lu, Wenpeng
    Pan, Yan
    Wu, Hao
    Wang, Rongyao
    Yu, Rui
    COMPUTERS & ELECTRICAL ENGINEERING, 2021, 90
  • [35] Retrospective Loss: Looking Back to Improve Training of Deep Neural Networks
    Jandial, Surgan
    Chopra, Ayush
    Sarkar, Mausoom
    Gupta, Piyush
    Krishnamurthy, Balaji
    Balasubramanian, Vineeth
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 1123 - 1131
  • [36] Robust Loss Functions under Label Noise for Deep Neural Networks
    Ghosh, Aritra
    Kumar, Himanshu
    Sastry, P. S.
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 1919 - 1925
  • [37] Jamming transition as a paradigm to understand the loss landscape of deep neural networks
    Geiger, Mario
    Spigler, Stefano
    d'Ascoli, Stephane
    Sagun, Levent
    Baity-Jesi, Marco
    Biroli, Giulio
    Wyart, Matthieu
    PHYSICAL REVIEW E, 2019, 100 (01)
  • [38] Defect Detection of Rail Surface with Deep Convolutional Neural Networks
    Liang, Zhicong
    Zhang, Hui
    Liu, Li
    He, Zhendong
    Zheng, Kai
    2018 13TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2018, : 1317 - 1322
  • [39] Assessment of Deep Convolutional Neural Networks for Road Surface Classification
    Nolte, Marcus
    Kister, Nikita
    Maurer, Markus
    2018 21ST INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2018, : 381 - 386
  • [40] Deep Convolutional Neural Networks for Detection of Rail Surface Defects
    Faghih-Roohi, Shahrzad
    Hajizadeh, Siamak
    Nunez, Alfredo
    Babuska, Robert
    De Schutter, Bart
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 2584 - 2589