CONTROLLING STOCHASTIC GRADIENT DESCENT USING STOCHASTIC APPROXIMATION FOR ROBUST DISTRIBUTED OPTIMIZATION

被引:0
|
作者
Jain, Adit [1 ]
Krishnamurthy, Vikram [1 ]
机构
[1] Cornell Univ, Ithaca, NY 14850 USA
基金
美国国家科学基金会;
关键词
Stochastic approximation; distributed optimization; Markov decision processes; POLICIES;
D O I
10.3934/naco.2024041
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This paper deals with the problem of controlling the stochastic gradient descent, performed by multiple learners where the aim is to estimate the respective arg min f using noisy gradients obtained by querying a stochastic oracle. Each query has a learning cost, and the noisy gradient response has varying degrees of noise variance, the bound of which is assumed to vary in a Markovian fashion. For a single learner, the decision problem is to choose when to query the oracle such that the learning cost is minimized. A constrained Markov decision process (CMDP) is formulated to solve the decision problem of a single learner. Structural results are proven for the optimal policy for the CMDP, which is shown to be threshold decreasing in the queue state. For multiple learners, a constrained switching control game is formulated for scheduling and controlling N learners querying the same oracle, one at a time. The structural results are extended for the optimal policy achieving the Nash equilibrium. The structural results are used to propose a stochastic approximation algorithm to search for the optimal policy, which tracks the parameters of the optimal policy using a sigmoidal approximation and does not require knowledge of the underlying transition probabilities. The paper also briefly discusses applications in federated learning and numerically shows the convergence properties of the proposed algorithm.
引用
收藏
页数:23
相关论文
共 50 条
  • [21] Convergence analysis of distributed stochastic gradient descent with shuffling
    Meng, Qi
    Chen, Wei
    Wang, Yue
    Ma, Zhi-Ming
    Liu, Tie-Yan
    NEUROCOMPUTING, 2019, 337 : 46 - 57
  • [22] Distributed and asynchronous Stochastic Gradient Descent with variance reduction
    Ming, Yuewei
    Zhao, Yawei
    Wu, Chengkun
    Li, Kuan
    Yin, Jianping
    NEUROCOMPUTING, 2018, 281 : 27 - 36
  • [23] Distributed Stochastic Gradient Descent With Compressed and Skipped Communication
    Phuong, Tran Thi
    Phong, Le Trieu
    Fukushima, Kazuhide
    IEEE ACCESS, 2023, 11 : 99836 - 99846
  • [24] Communication-Censored Distributed Stochastic Gradient Descent
    Li, Weiyu
    Wu, Zhaoxian
    Chen, Tianyi
    Li, Liping
    Ling, Qing
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (11) : 6831 - 6843
  • [25] Using the Stochastic Gradient Descent Optimization Algorithm on Estimating of Reactivity Ratios
    Fazakas-Anca, Iosif Sorin
    Modrea, Arina
    Vlase, Sorin
    MATERIALS, 2021, 14 (16)
  • [26] STOCHASTIC GRADIENT DESCENT ALGORITHM FOR STOCHASTIC OPTIMIZATION IN SOLVING ANALYTIC CONTINUATION PROBLEMS
    Bao, Feng
    Maier, Thomas
    FOUNDATIONS OF DATA SCIENCE, 2020, 2 (01): : 1 - 17
  • [27] ON UNIFORM-IN-TIME DIFFUSION APPROXIMATION FOR STOCHASTIC GRADIENT DESCENT
    Li, Lei
    Wang, Yuliang
    METHODS AND APPLICATIONS OF ANALYSIS, 2023, 30 (03) : 95 - 112
  • [28] ON THE PRIVACY OF NOISY STOCHASTIC GRADIENT DESCENT FOR CONVEX OPTIMIZATION
    Altschuler, Jason M.
    Bok, Jinho
    Talwar, Kunal
    SIAM JOURNAL ON COMPUTING, 2024, 53 (04) : 969 - 1001
  • [29] Comparison of the Stochastic Gradient Descent Based Optimization Techniques
    Yazan, Ersan
    Talu, M. Fatih
    2017 INTERNATIONAL ARTIFICIAL INTELLIGENCE AND DATA PROCESSING SYMPOSIUM (IDAP), 2017,
  • [30] BYZANTINE-ROBUST STOCHASTIC GRADIENT DESCENT FOR DISTRIBUTED LOW-RANK MATRIX COMPLETION
    He, Xuechao
    Ling, Qing
    Chen, Tianyi
    2019 IEEE DATA SCIENCE WORKSHOP (DSW), 2019, : 322 - 326