Bias of Homotopic Gradient Descent for the Hinge Loss

被引:0
|
作者
Denali Molitor
Deanna Needell
Rachel Ward
机构
[1] University of California,Department of Mathematics
[2] Los Angeles,Department of Mathematics
[3] University of Texas at Austin,undefined
来源
关键词
Convex optimization; Gradient descent; Support vector machines; Hinge loss; Non-smooth optimization;
D O I
暂无
中图分类号
学科分类号
摘要
Gradient descent is a simple and widely used optimization method for machine learning. For homogeneous linear classifiers applied to separable data, gradient descent has been shown to converge to the maximal-margin (or equivalently, the minimal-norm) solution for various smooth loss functions. The previous theory does not, however, apply to the non-smooth hinge loss which is widely used in practice. Here, we study the convergence of a homotopic variant of gradient descent applied to the hinge loss and provide explicit convergence rates to the maximal-margin solution for linearly separable data.
引用
收藏
页码:621 / 647
页数:26
相关论文
共 50 条
  • [1] Bias of Homotopic Gradient Descent for the Hinge Loss
    Molitor, Denali
    Needell, Deanna
    Ward, Rachel
    APPLIED MATHEMATICS AND OPTIMIZATION, 2021, 84 (01): : 621 - 647
  • [2] Epoch Gradient Descent for Smoothed Hinge-loss Linear SVMs
    Lee, Soomin
    Nedic, Angelia
    2013 AMERICAN CONTROL CONFERENCE (ACC), 2013, : 4789 - 4794
  • [3] Hinge Classification Algorithm Based on Asynchronous Gradient Descent
    Yan, Xiaodan
    Zhang, Tianxin
    Cui, Baojiang
    Deng, Jiangdong
    ADVANCES ON BROAD-BAND WIRELESS COMPUTING, COMMUNICATION AND APPLICATIONS, BWCCA-2017, 2018, 12 : 459 - 468
  • [4] The implicit bias of gradient descent on nonseparable data
    Ji, Ziwei
    Telgarsky, Matus
    CONFERENCE ON LEARNING THEORY, VOL 99, 2019, 99
  • [5] The Implicit Bias of Gradient Descent on Separable Data
    Soudry, Daniel
    Hoffer, Elad
    Nacson, Mor Shpigel
    Gunasekar, Suriya
    Srebro, Nathan
    JOURNAL OF MACHINE LEARNING RESEARCH, 2018, 19
  • [6] On the Implicit Bias of Gradient Descent for Temporal Extrapolation
    Cohen-Karlik, Edo
    Ben David, Avichai
    Cohen, Nadav
    Globerson, Amir
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [7] The implicit bias of gradient descent on separable data
    Soudry, Daniel
    Hoffer, Elad
    Nacson, Mor Shpigel
    Gunasekar, Suriya
    Srebro, Nathan
    Journal of Machine Learning Research, 2018, 19
  • [8] Implicit Bias of Gradient Descent on Reparametrized Models: On Equivalence to Mirror Descent
    Li, Zhiyuan
    Wang, Tianhao
    Lee, Jason D.
    Arora, Sanjeev
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [9] Implicit Bias of Gradient Descent on Linear Convolutional Networks
    Gunasekar, Suriya
    Lee, Jason D.
    Soudry, Daniel
    Srebro, Nathan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [10] Learning a Single Neuron with Bias Using Gradient Descent
    Vardi, Gal
    Yehudai, Gilad
    Shamir, Ohad
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34