SALR: Sharpness-Aware Learning Rate Scheduler for Improved Generalization

被引:0
|
作者
Yue, Xubo [1 ]
Nouiehed, Maher [2 ]
Al Kontar, Raed [1 ]
机构
[1] Univ Michigan, Dept Ind & Operat Engn, Ann Arbor, MI 48109 USA
[2] Amer Univ Beirut, Dept Ind Engn & Management, Beirut 1072020, Lebanon
基金
美国国家科学基金会;
关键词
Schedules; Deep learning; Neural networks; Convergence; Bayes methods; Training; Stochastic processes; generalization; learning rate schedule; sharpness;
D O I
10.1109/TNNLS.2023.3263393
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In an effort to improve generalization in deep learning and automate the process of learning rate scheduling, we propose SALR: a sharpness-aware learning rate update technique designed to recover flat minimizers. Our method dynamically updates the learning rate of gradient-based optimizers based on the local sharpness of the loss function. This allows optimizers to automatically increase learning rates at sharp valleys to increase the chance of escaping them. We demonstrate the effectiveness of SALR when adopted by various algorithms over a broad range of networks. Our experiments indicate that SALR improves generalization, converges faster, and drives solutions to significantly flatter regions.
引用
收藏
页码:12518 / 12527
页数:10
相关论文
共 50 条
  • [1] Implicit Sharpness-Aware Minimization for Domain Generalization
    Dong, Mingrong
    Yang, Yixuan
    Zeng, Kai
    Wang, Qingwang
    Shen, Tao
    REMOTE SENSING, 2024, 16 (16)
  • [2] Sharpness-Aware Gradient Matching for Domain Generalization
    Wang, Pengfei
    Zhang, Zhaoxiang
    Lei, Zhen
    Zhang, Lei
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 3769 - 3778
  • [3] Sharpness-Aware Minimization Improves Language Model Generalization
    Bahri, Dara
    Mobahi, Hossein
    Tay, Yi
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 7360 - 7371
  • [4] Sharpness-Aware Lookahead for Accelerating Convergence and Improving Generalization
    Tan, Chengli
    Zhang, Jiangshe
    Liu, Junmin
    Gong, Yihong
    IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, 46 (12) : 10375 - 10388
  • [5] Adaptive Sharpness-Aware Minimization for Adversarial Domain Generalization
    Xie, Tianci
    Li, Tao
    Wu, Ruoxue
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024,
  • [6] FedGAMMA: Federated Learning With Global Sharpness-Aware Minimization
    Dai, Rong
    Yang, Xun
    Sun, Yan
    Shen, Li
    Tian, Xinmei
    Wang, Meng
    Zhang, Yongdong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 35 (12) : 1 - 14
  • [7] Research on the improvement of domain generalization by the fusion of invariant features and sharpness-aware minimization
    Yixuan Yang
    Mingrong Dong
    Kai Zeng
    Tao Shen
    The Journal of Supercomputing, 2025, 81 (1)
  • [8] AdaSAM: Boosting sharpness-aware minimization with adaptive learning rate and momentum for neural networks
    Sun, Hao
    Shen, Li
    Zhong, Qihuang
    Ding, Liang
    Chen, Shixiang
    Sun, Jingwei
    Li, Jing
    Sun, Guangzhong
    Tao, Dacheng
    NEURAL NETWORKS, 2024, 169 : 506 - 519
  • [9] Random Sharpness-Aware Minimization
    Liu, Yong
    Mai, Siqi
    Cheng, Minhao
    Chen, Xiangning
    Hsieh, Cho-Jui
    You, Yang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [10] TOWARDS GENERALIZATION OF MEDICAL IMAGING AI MODELS: SHARPNESS-AWARE MINIMIZERS AND BEYOND
    Anand, Deepa
    Patil, Rohan
    Agrawal, Utkarsh
    Rahul, V
    Ravishankar, Hariharan
    Sudhakar, Prasad
    2022 IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (IEEE ISBI 2022), 2022,