Adaptive Stochastic Gradient Descent (SGD) for erratic datasets

被引:2
|
作者
Dagal, Idriss [1 ]
Tanrioven, Kursat [1 ]
Nayir, Ahmet [1 ]
Akin, Burak [2 ]
机构
[1] Istanbul Beykent Univ, Elect Engn, Hadim Koruyolu Caddesi 19, TR-34450 Istanbul, Turkiye
[2] Yildiz Tech Univ, Elect Engn, Davutpasa Caddesi, TR-34220 Istanbul, Turkiye
关键词
Gradient descent; Stochastic Gradient Descent; Accuracy; Principal Component Analysis; QUASI-NEWTON METHOD; NEURAL NETWORKS; ALGORITHM; MLP;
D O I
10.1016/j.future.2024.107682
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Stochastic Gradient Descent (SGD) is a highly efficient optimization algorithm, particularly well suited for large datasets due to its incremental parameter updates. In this study, we apply SGD to a simple linear classifier using logistic regression, a widely used method for binary classification tasks. Unlike traditional batch Gradient Descent (GD), which processes the entire dataset simultaneously, SGD offers enhanced scalability and performance for streaming and large-scale data. Our experiments reveal that SGD outperforms GD across multiple performance metrics, achieving 45.83% accuracy compared to GD's 41.67 %, and excelling in precision (60 % vs. 45.45 %), recall (100 % vs. 60 %), and F1-score (100 % vs. 62 %). Additionally, SGD achieves 99.99 % of Principal Component Analysis (PCA) accuracy, slightly surpassing GD's 99.92 %. These results highlight SGD's superior efficiency and flexibility for large-scale data environments, driven by its ability to balance precision and recall effectively. To further enhance SGD's robustness, the proposed method incorporates adaptive learning rates, momentum, and logistic regression, addressing traditional GD drawbacks. These modifications improve the algorithm's stability, convergence behavior, and applicability to complex, large-scale optimization tasks where standard GD often struggles, making SGD a highly effective solution for challenging data-driven scenarios.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Cervical Cancer Detection Using Ensemble Neural Network Algorithm with Stochastic Gradient Descent (SGD) Optimizer
    K. Shanthi
    S. Manimekalai
    SN Computer Science, 5 (8)
  • [32] Adaptive Gradient Estimation Stochastic Parallel Gradient Descent Algorithm for Laser Beam Cleanup
    Ma, Shiqing
    Yang, Ping
    Lai, Boheng
    Su, Chunxuan
    Zhao, Wang
    Yang, Kangjian
    Jin, Ruiyan
    Cheng, Tao
    Xu, Bing
    PHOTONICS, 2021, 8 (05)
  • [33] Entropy-SGD: biasing gradient descent into wide valleys
    Chaudhari, Pratik
    Choromanska, Anna
    Soatto, Stefano
    LeCun, Yann
    Baldassi, Carlo
    Borgs, Christian
    Chayes, Jennifer
    Sagun, Levent
    Zecchina, Riccardo
    JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2019, 2019 (12):
  • [34] Adaptive stochastic parallel gradient descent approach for efficient fiber coupling
    Hu, Qintao
    Zhen, Liangli
    Yao, Mao
    Zhu, Shiwei
    Zhou, Xi
    Zhou, Guozhong
    OPTICS EXPRESS, 2020, 28 (09) : 13141 - 13154
  • [35] Adaptive Distributed Stochastic Gradient Descent for Minimizing Delay in the Presence of Stragglers
    Hanna, Serge Kas
    Bitar, Rawad
    Parag, Parimal
    Dasari, Venkat
    El Rouayheb, Salim
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 4262 - 4266
  • [36] Parallel Fractional Stochastic Gradient Descent With Adaptive Learning for Recommender Systems
    Elahi, Fatemeh
    Fazlali, Mahmood
    Malazi, Hadi Tabatabaee
    Elahi, Mehdi
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (03) : 470 - 483
  • [37] MindTheStep-AsyncPSGD: Adaptive Asynchronous Parallel Stochastic Gradient Descent
    Backstrom, Karl
    Papatriantafilou, Marina
    Tsigas, Philippas
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 16 - 25
  • [38] aSGD: Stochastic Gradient Descent with Adaptive Batch Size for Every Parameter
    Shi, Haoze
    Yang, Naisen
    Tang, Hong
    Yang, Xin
    MATHEMATICS, 2022, 10 (06)