A stochastic gradient descent algorithm for structural risk minimisation

被引:0
|
作者
Ratsaby, J [1 ]
机构
[1] UCL, London WC1E 6BT, England
来源
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Structural risk minimisation (SRM) is a general complexity regularization method which automatically selects the model complexity that approximately minimises the misclassification error probability of the empirical risk minimiser. It does so by adding a complexity penalty term epsilon(m, k) to the empirical risk of the candidate hypotheses and then for any fixed sample size m it minimises the sum with respect to the model complexity variable k. When learning multicategory classification there are M subsamples m(i), corresponding to the M pattern classes with a priori probabilities p(i), 1 less than or equal to i less than or equal to M. Using the usual representation for a multi-category classifier as M individual boolean classifiers, the penalty becomes Sigma(i=1)(M) P(i)epsilon(m(i), k(i)). If the m(i) are given then the standard SRM trivially applies here by minimizing the penalised empirical risk with respect to k(i),1,..., M. However, in situations where the total sample size Sigma(i=1)(M) m(i), needs to be minimal one needs to also minimize the penalised empirical risk with respect to the variables mi, i = 1,..., M. The obvious problem is that the empirical risk can only be defined after the subsamples (and hence their sizes) are given (known). Utilising an on-line stochastic gradient descent approach, this paper overcomes this difficulty and introduces a sample-querying algorithm which extends the standard SRM principle. It minimises the penalised empirical risk not only with respect to the ki, as the standard SRM does, but also with respect to the m(i,) i = 1,...,M. The challenge here is in defining a stochastic empirical criterion which when minimised yields a sequence of subsample-size vectors which asymptotically achieve the Bayes' optimal error convergence rate.
引用
收藏
页码:205 / 220
页数:16
相关论文
共 50 条
  • [21] Stochastic parallel gradient descent algorithm for adaptive optics system
    Ma H.
    Zhang P.
    Zhang J.
    Fan C.
    Wang Y.
    Qiangjiguang Yu Lizishu/High Power Laser and Particle Beams, 2010, 22 (06): : 1206 - 1210
  • [22] A Novel Stochastic Gradient Descent Algorithm for Learning Principal Subspaces
    Le Lan, Charline
    Greaves, Joshua
    Farebrother, Jesse
    Rowland, Mark
    Pedregosa, Fabian
    Agarwal, Rishabh
    Bellemare, Marc
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 206, 2023, 206
  • [23] Comparing Stochastic Gradient Descent and Mini-batch Gradient Descent Algorithms in Loan Risk Assessment
    Adigun, Abodunrin AbdulGafar
    Yinka-Banjo, Chika
    INFORMATICS AND INTELLIGENT APPLICATIONS, 2022, 1547 : 283 - 296
  • [24] Empirical Risk Minimization and Stochastic Gradient Descent for Relational Data
    Veitch, Vicor
    Austern, Morgane
    Zhou, Wenda
    Blei, David M.
    Orbanz, Peter
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [25] hr Dimension independent excess risk by stochastic gradient descent
    Chen, Xi
    Liu, Qiang
    Tong, Xin T.
    ELECTRONIC JOURNAL OF STATISTICS, 2022, 16 (02): : 4547 - 4603
  • [26] Risk optimization using the Chernoff bound and stochastic gradient descent
    Carlon, Andre Gustavo
    Kroetz, Henrique Machado
    Torii, Andre Jacomel
    Lopez, Rafael Holdorf
    Fadel Miguel, Leandro Fleck
    RELIABILITY ENGINEERING & SYSTEM SAFETY, 2022, 223
  • [27] Adaptive Gradient Estimation Stochastic Parallel Gradient Descent Algorithm for Laser Beam Cleanup
    Ma, Shiqing
    Yang, Ping
    Lai, Boheng
    Su, Chunxuan
    Zhao, Wang
    Yang, Kangjian
    Jin, Ruiyan
    Cheng, Tao
    Xu, Bing
    PHOTONICS, 2021, 8 (05)
  • [28] Unforgeability in Stochastic Gradient Descent
    Baluta, Teodora
    Nikolic, Ivica
    Jain, Racchit
    Aggarwal, Divesh
    Saxena, Prateek
    PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 1138 - 1152
  • [29] Preconditioned Stochastic Gradient Descent
    Li, Xi-Lin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (05) : 1454 - 1466
  • [30] Stochastic Reweighted Gradient Descent
    El Hanchi, Ayoub
    Stephens, David A.
    Maddison, Chris J.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,