Adaptive Negative Curvature Descent with Applications in Non-convex Optimization

被引:0
|
作者
Liu, Mingrui [1 ]
Li, Zhe [1 ]
Wang, Xiaoyu [2 ]
Yi, Jinfeng [3 ]
Yang, Tianbao [1 ]
机构
[1] Univ Iowa, Dept Comp Sci, Iowa City, IA 52242 USA
[2] Intellifusion, Parlin, NJ USA
[3] JD AI Res, Stanford, CA USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Negative curvature descent (NCD) method has been utilized to design deterministic or stochastic algorithms for non-convex optimization aiming at finding second-order stationary points or local minima In existing studies, NCD needs to approximate the smallest eigen-value of the Hessian matrix with a sufficient precision (e.g., epsilon(2) << 1) in order to achieve a sufficiently accurate second-order stationary solution (i.e., lambda(min)(del(2)f(x)) >= -epsilon(2)). One issue with this approach is that the target precision epsilon(2) is usually set to be very small in order to find a high quality solution, which increases the complexity for computing a negative curvature. To address this issue, we propose an adaptive NCD to allow an adaptive error dependent on the current gradient's magnitude in approximating the smallest eigen-value of the Hessian, and to encourage competition between a noisy NCD step and gradient descent step. We consider the applications of the proposed adaptive NCD for both deterministic and stochastic non-convex optimization, and demonstrate that it can help reduce the the overall complexity in computing the negative curvatures during the course of optimization without sacrificing the iteration complexity.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Projected Gradient Descent for Non-Convex Sparse Spike Estimation
    Traonmilin, Yann
    Aujol, Jean-Francois
    Leclaire, Arthur
    IEEE SIGNAL PROCESSING LETTERS, 2020, 27 : 1110 - 1114
  • [32] Fast Block Coordinate Descent for Non-Convex Group Regularizations
    Ida, Yasutoshi
    Kanai, Sekitoshi
    Kumagai, Atsutoshi
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 206, 2023, 206
  • [33] Generalization Bound of Gradient Descent for Non-Convex Metric Learning
    Dong, Mingzhi
    Yang, Xiaochen
    Zhu, Rui
    Wang, Yujiang
    Xue, Jing-Hao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [34] Scaling up stochastic gradient descent for non-convex optimisation
    Mohamad, Saad
    Alamri, Hamad
    Bouchachia, Abdelhamid
    MACHINE LEARNING, 2022, 111 (11) : 4039 - 4079
  • [35] Damping proximal coordinate descent algorithm for non-convex regularization
    Pan, Zheng
    Lin, Ming
    Hou, Guangdong
    Zhang, Changshui
    NEUROCOMPUTING, 2015, 152 : 151 - 163
  • [36] Scaling up stochastic gradient descent for non-convex optimisation
    Saad Mohamad
    Hamad Alamri
    Abdelhamid Bouchachia
    Machine Learning, 2022, 111 : 4039 - 4079
  • [37] On the Convergence of (Stochastic) Gradient Descent with Extrapolation for Non-Convex Minimization
    Xu, Yi
    Yuan, Zhuoning
    Yang, Sen
    Jin, Rong
    Yang, Tianbao
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4003 - 4009
  • [38] Robust Adaptive Beamforming Based on Non-convex Quadratic Optimization with Semidefinite Relaxation
    Chen, Yong
    Wang, Fang
    2018 IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING, COMMUNICATIONS AND COMPUTING (ICSPCC), 2018,
  • [39] Decentralized non-convex online optimization with adaptive momentum estimation and quantized communication
    Lv, Yunshan
    Xiong, Hailing
    Zhang, Fuqing
    Dong, Shengying
    Dai, Xiangguang
    COMPLEX & INTELLIGENT SYSTEMS, 2025, 11 (04)
  • [40] Adaptive Rooted Tree Optimization for Non-convex Economic Dispatch Cost Function
    Edagbami, Sola E.
    Yinka-Banjo, Chika
    Uwadia, Charles O.
    2021 IEEE PES/IAS POWERAFRICA CONFERENCE, 2021, : 609 - 613