KALMAN OPTIMIZER FOR CONSISTENT GRADIENT DESCENT

被引:3
|
作者
Yang, Xingyi [1 ]
机构
[1] Univ Calif San Diego, Dept Elect & Comp Engn, 9500 Gilman Dr, La Jolla, CA 92093 USA
关键词
Stochastic gradient descent; Kalman Filtering; Optimization;
D O I
10.1109/ICASSP39728.2021.9414588
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Deep neural networks (DNN) are typically optimized using stochastic gradient descent (SGD). However, the estimation of the gradient using stochastic samples tends to be noisy and unreliable, resulting in large gradient variance and bad convergence. In this paper, we propose Kalman Optimizor (KO), an efficient stochastic optimization algorithm that adopts Kalman filter to make consistent estimation of the local gradient by solving an adaptive filtering problem. Our method reduces estimation variance in stochastic gradient descent by incorporating the historic state of the optimization. It aims to improve noisy gradient direction as well as accelerate the convergence of learning. We demonstrate the effectiveness of the proposed Kalman Optimizer under various optimization tasks where it is shown to achieve superior and robust performance. The code is available at https: //github.com/Adamdad/Filter-Gradient-Decent.
引用
收藏
页码:3900 / 3904
页数:5
相关论文
共 50 条
  • [1] Gradient Descent: The Ultimate Optimizer
    Chandra, Kartik
    Xie, Audrey
    Ragan-Kelley, Jonathan
    Meijer, Erik
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [2] A Cost-based Optimizer for Gradient Descent Optimization
    Kaoudi, Zoi
    Quiane-Ruiz, Jorge-Arnulfo
    Thirumuruganathan, Saravanan
    Chawla, Sanjay
    Agrawal, Divy
    [J]. SIGMOD'17: PROCEEDINGS OF THE 2017 ACM INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA, 2017, : 977 - 992
  • [3] Fractional Gradient Descent Optimizer for Linear Classifier Support Vector Machine
    Hapsari, Dian Puspita
    Utoyo, Imam
    Purnami, Santi Wulan
    [J]. 2020 THIRD INTERNATIONAL CONFERENCE ON VOCATIONAL EDUCATION AND ELECTRICAL ENGINEERING (ICVEE): STRENGTHENING THE FRAMEWORK OF SOCIETY 5.0 THROUGH INNOVATIONS IN EDUCATION, ELECTRICAL, ENGINEERING AND INFORMATICS ENGINEERING, 2020,
  • [4] Hybrid Gradient Descent Grey Wolf Optimizer for Optimal Feature Selection
    Kitonyi, Peter Mule
    Segera, Davies Rene
    [J]. BIOMED RESEARCH INTERNATIONAL, 2021, 2021
  • [5] Tangent-cut optimizer on gradient descent: an approach towards Hybrid Heuristics
    Saptarshi Biswas
    Subhrapratim Nath
    Sumagna Dey
    Utsha Majumdar
    [J]. Artificial Intelligence Review, 2022, 55 : 1121 - 1147
  • [6] Modified Convolutional Neural Network Based on Dropout and the Stochastic Gradient Descent Optimizer
    Yang, Jing
    Yang, Guanci
    [J]. ALGORITHMS, 2018, 11 (03):
  • [7] Tangent-cut optimizer on gradient descent: an approach towards Hybrid Heuristics
    Biswas, Saptarshi
    Nath, Subhrapratim
    Dey, Sumagna
    Majumdar, Utsha
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (02) : 1121 - 1147
  • [8] Smartphone Orientation Estimation Algorithm Combining Kalman Filter With Gradient Descent
    Yean, Seanglidet
    Lee, Bu Sung
    Yeo, Chai Kiat
    Vun, Chan Hua
    Oh, Hong Lye
    [J]. IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2018, 22 (05) : 1421 - 1433
  • [9] An enhanced learning algorithm with a particle filter-based gradient descent optimizer method
    Patcharin Kamsing
    Peerapong Torteeka
    Soemsak Yooyen
    [J]. Neural Computing and Applications, 2020, 32 : 12789 - 12800
  • [10] An enhanced learning algorithm with a particle filter-based gradient descent optimizer method
    Kamsing, Patcharin
    Torteeka, Peerapong
    Yooyen, Soemsak
    [J]. NEURAL COMPUTING & APPLICATIONS, 2020, 32 (16): : 12789 - 12800