Accelerating Gradient Descent with Projective Response Surface Methodology

被引:3
|
作者
Senov, Alexander [1 ]
机构
[1] St Petersburg State Univ, Fac Math & Mech, Univ Sky Prospekt 28, St Petersburg 198504, Russia
基金
俄罗斯科学基金会;
关键词
Least-squares; Steepest descent; Quadratic programming; Projective methods;
D O I
10.1007/978-3-319-69404-7_34
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a new modification of gradient descent algorithm based on surrogate optimization with projection into low-dimensional space. It consequently builds an approximation of the target function in low-dimensional space and takes the approximation optimum point mapped back to original parameter space as the next parameter estimate. An additional projection step is used to fight the curse of dimensionality. Major advantage of the proposed modification is that it does not change gradient descent iterations, thus it may be used with almost any zero- or first-order iterative method. We give a theoretical motivation for the proposed algorithm and experimentally illustrate its properties on modelled data.
引用
收藏
页码:376 / 382
页数:7
相关论文
共 50 条
  • [1] Projective Fisher Information for Natural Gradient Descent
    Kaul, Piyush
    Lall, Brejesh
    [J]. IEEE Transactions on Artificial Intelligence, 2023, 4 (02): : 304 - 314
  • [2] Projective Approximation Based Gradient Descent Modification
    Senov, Alexander
    Granichin, Oleg
    [J]. IFAC PAPERSONLINE, 2017, 50 (01): : 3899 - 3904
  • [3] Accelerating Federated Learning via Momentum Gradient Descent
    Liu, Wei
    Chen, Li
    Chen, Yunfei
    Zhang, Wenyi
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2020, 31 (08) : 1754 - 1766
  • [4] Accelerating gradient descent and Adam via fractional gradients
    Shin, Yeonjong
    Darbon, Jerome
    Karniadakis, George Em
    [J]. NEURAL NETWORKS, 2023, 161 : 185 - 201
  • [5] Lightweight Projective Derivative Codes for Compressed Asynchronous Gradient Descent
    Soto, Pedro
    Ilmer, Ilia
    Guan, Haibin
    Li, Jun
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [6] Accelerating Stochastic Gradient Descent Based Matrix Factorization on FPGA
    Zhou, Shijie
    Kannan, Rajgopal
    Prasanna, Viktor K.
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2020, 31 (08) : 1897 - 1911
  • [7] Accelerating Asynchronous Stochastic Gradient Descent for Neural Machine Translation
    Bogoychev, Nikolay
    Junczys-Dowmunt, Marcin
    Heafield, Kenneth
    Aji, Alham Fikri
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 2991 - 2996
  • [8] Accelerating Rescaled Gradient Descent: Fast Optimization of Smooth Functions
    Wilson, Ashia C.
    Mackey, Lester
    Wibisono, Andre
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] Accelerating Minibatch Stochastic Gradient Descent Using Typicality Sampling
    Peng, Xinyu
    Li, Li
    Wang, Fei-Yue
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (11) : 4649 - 4659
  • [10] Accelerating deep neural network training with inconsistent stochastic gradient descent
    Wang, Linnan
    Yang, Yi
    Min, Renqiang
    Chakradhar, Srimat
    [J]. NEURAL NETWORKS, 2017, 93 : 219 - 229