Superpolynomial Lower Bounds for Learning One-Layer Neural Networks using Gradient Descent

被引:0
|
作者
Goel, Surbhi [1 ]
Gollakota, Aravind [1 ]
Jin, Zhihan [2 ]
Karmalkar, Sushrut [1 ]
Klivans, Adam [1 ]
机构
[1] Univ Texas Austin, Dept Comp Sci, Austin, TX 78712 USA
[2] Shanghai Jiao Tong Univ, Dept Comp Sci, Shanghai, Peoples R China
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We give the first superpolynomial lower bounds for learning one-layer neural networks with respect to the Gaussian distribution for a broad class of algorithms. We prove that gradient descent run on any classifier with respect to square loss will fail to achieve small test error in polynomial time. Prior work held only for gradient descent run with small batch sizes and sufficiently smooth classifiers. For classification, we give a stronger result, namely that any statistical query (SQ) algorithm (including gradient descent) will fail to achieve small test error in polynomial time. Our lower bounds hold for commonly used activations such as ReLU and sigmoid. The core of our result relies on a novel construction of a simple family of neural networks that are exactly orthogonal with respect to all spherically symmetric distributions.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Superpolynomial Lower Bounds for Learning One-Layer Neural Networks using Gradient Descent
    Goel, Surbhi
    Gollakota, Aravind
    Jin, Zhihan
    Karmalkar, Sushrut
    Klivans, Adam
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [2] Gradient Descent for One-Hidden-Layer Neural Networks: Polynomial Convergence and SQ Lower Bounds
    Vempala, Santosh
    Wilmes, John
    [J]. CONFERENCE ON LEARNING THEORY, VOL 99, 2019, 99
  • [3] Minimax Lower Bounds for Transfer Learning with Linear and One-hidden Layer Neural Networks
    Kalan, Seyed Mohammadreza Mousavi
    Fabian, Zalan
    Avestimehr, Salman
    Soltanolkotabi, Mahdi
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [4] An effective and efficient green federated learning method for one-layer neural networks
    Fontenla-Romero, Oscar
    Guijarro-Berdinas, Bertha
    Hernandez-Pereira, Elena
    Perez-Sanchez, Beatriz
    [J]. 39TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2024, 2024, : 1050 - 1052
  • [5] FedHEONN: Federated and homomorphically encrypted learning method for one-layer neural networks
    Fontenla-Romero, Oscar
    Guijarro-Berdinas, Bertha
    Hernandez-Pereira, Elena
    Perez-Sanchez, Beatriz
    [J]. FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2023, 149 : 200 - 211
  • [6] A global optimum approach for one-layer neural networks
    Castillo, E
    Fontenla-Romero, O
    Guijarro-Berdiñas, B
    Alonso-Betanzos, A
    [J]. NEURAL COMPUTATION, 2002, 14 (06) : 1429 - 1449
  • [7] Learning One-hidden-layer ReLU Networks via Gradient Descent
    Zhang, Xiao
    Yu, Yaodong
    Wang, Lingxiao
    Gu, Quanquan
    [J]. 22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [8] Learning Distributions Generated by One-Layer ReLU Networks
    Wu, Shanshan
    Dimakis, Alexandros G.
    Sanghavi, Sujay
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] An incremental non-iterative learning method for one-layer feedforward neural networks
    Fontenla-Romero, Oscar
    Perez-Sanchez, Beatriz
    Guijarro-Berdinas, Bertha
    [J]. APPLIED SOFT COMPUTING, 2018, 70 : 951 - 958
  • [10] Regularized One-Layer Neural Networks for Distributed and Incremental Environments
    Fontenla-Romero, Oscar
    Guijarro-Berdinas, Bertha
    Perez-Sanchez, Beatriz
    [J]. ADVANCES IN COMPUTATIONAL INTELLIGENCE (IWANN 2021), PT II, 2021, 12862 : 343 - 355