Optimal stochastic gradient descent algorithm for filtering

被引:0
|
作者
Turali, M. Yigit [1 ]
Koc, Ali T. [1 ]
Kozat, Suleyman S. [1 ]
机构
[1] Bilkent Univ, Dept Elect & Elect Engn, TR-06800 Ankara, Turkiye
关键词
Learning rate; Linear filtering; Optimization; Stochastic gradient descent; PREDICTION;
D O I
10.1016/j.dsp.2024.104731
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Stochastic Gradient Descent (SGD) is a fundamental optimization technique in machine learning, due to its efficiency in handling large-scale data. Unlike typical SGD applications, which rely on stochastic approximations, this work explores the convergence properties of SGD from a deterministic perspective. We address the crucial aspect of learning rate settings, a common obstacle in optimizing SGD performance, particularly in complex environments. In contrast to traditional methods that often provide convergence results based on statistical expectations (which are usually not justified), our approach introduces universally applicable learning rates. These rates ensure that a model trained with SGD matches the performance of the best linear filter asymptotically, applicable irrespective of the data sequence length and independent of statistical assumptions about the data. By establishing learning rates that scale as mu = O(1/t), we offer a solution that sidesteps the need for prior data knowledge, a prevalent limitation in real-world applications. To this end, we provide a robust framework for SGD's application across varied settings, guaranteeing convergence results that hold under both deterministic and stochastic scenarios without any underlying assumptions.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] A stochastic multiple gradient descent algorithm
    Mercier, Quentin
    Poirion, Fabrice
    Desideri, Jean-Antoine
    [J]. EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2018, 271 (03) : 808 - 817
  • [2] A Stochastic Gradient Descent Approach for Stochastic Optimal Control
    Archibald, Richard
    Bao, Feng
    Yong, Jiongmin
    [J]. EAST ASIAN JOURNAL ON APPLIED MATHEMATICS, 2020, 10 (04) : 635 - 658
  • [3] Sign Based Derivative Filtering for Stochastic Gradient Descent
    Berestizshevsky, Konstantin
    Even, Guy
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: DEEP LEARNING, PT II, 2019, 11728 : 208 - 219
  • [4] Learning Stochastic Optimal Policies via Gradient Descent
    Massaroli, Stefano
    Poli, Michael
    Peluchetti, Stefano
    Park, Jinkyoo
    Yamashita, Atsushi
    Asama, Hajime
    [J]. IEEE CONTROL SYSTEMS LETTERS, 2022, 6 : 1094 - 1099
  • [5] A new stochastic gradient descent possibilistic clustering algorithm
    Koutsimpela, Angeliki
    Koutroumbas, Konstantinos D.
    [J]. AI COMMUNICATIONS, 2022, 35 (02) : 47 - 64
  • [6] Fast Convergence Stochastic Parallel Gradient Descent Algorithm
    Hu Dongting
    Shen Wen
    Ma Wenchao
    Liu Xinyu
    Su Zhouping
    Zhu Huaxin
    Zhang Xiumei
    Que Lizhi
    Zhu Zhuowei
    Zhang Yixin
    Chen Guoqing
    Hu Lifa
    [J]. LASER & OPTOELECTRONICS PROGRESS, 2019, 56 (12)
  • [7] The Improved Stochastic Fractional Order Gradient Descent Algorithm
    Yang, Yang
    Mo, Lipo
    Hu, Yusen
    Long, Fei
    [J]. FRACTAL AND FRACTIONAL, 2023, 7 (08)
  • [8] Guided Stochastic Gradient Descent Algorithm for inconsistent datasets
    Sharma, Anuraganand
    [J]. APPLIED SOFT COMPUTING, 2018, 73 : 1068 - 1080
  • [9] Stochastic Approximate Gradient Descent via the Langevin Algorithm
    Qiu, Yixuan
    Wang, Xiao
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 5428 - 5435
  • [10] A stochastic gradient descent algorithm for structural risk minimisation
    Ratsaby, J
    [J]. ALGORITHMIC LEARNING THEORY, PROCEEDINGS, 2003, 2842 : 205 - 220