Relaxation Subgradient Algorithms with Machine Learning Procedures

被引:7
|
作者
Krutikov, Vladimir [1 ]
Gutova, Svetlana [1 ]
Tovbis, Elena [2 ]
Kazakovtsev, Lev [2 ]
Semenkin, Eugene [2 ]
机构
[1] Kemerovo State Univ, Dept Appl Math, Krasnaya St 6, Kemerovo 650043, Russia
[2] Reshetnev Siberian State Univ Sci & Technol, Inst Informat & Telecommun, Prosp Krasnoyarskiy Rabochiy 31, Krasnoyarsk 660031, Russia
关键词
relaxation subgradient methods; space dilation; nonsmooth minimization methods; machine learning algorithm; APPROXIMATION; MINIMIZATION; SELECTION;
D O I
10.3390/math10213959
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
In the modern digital economy, optimal decision support systems, as well as machine learning systems, are becoming an integral part of production processes. Artificial neural network training as well as other engineering problems generate such problems of high dimension that are difficult to solve with traditional gradient or conjugate gradient methods. Relaxation subgradient minimization methods (RSMMs) construct a descent direction that forms an obtuse angle with all subgradients of the current minimum neighborhood, which reduces to the problem of solving systems of inequalities. Having formalized the model and taking into account the specific features of subgradient sets, we reduced the problem of solving a system of inequalities to an approximation problem and obtained an efficient rapidly converging iterative learning algorithm for finding the direction of descent, conceptually similar to the iterative least squares method. The new algorithm is theoretically substantiated, and an estimate of its convergence rate is obtained depending on the parameters of the subgradient set. On this basis, we have developed and substantiated a new RSMM, which has the properties of the conjugate gradient method on quadratic functions. We have developed a practically realizable version of the minimization algorithm that uses a rough one-dimensional search. A computational experiment on complex functions in a space of high dimension confirms the effectiveness of the proposed algorithm. In the problems of training neural network models, where it is required to remove insignificant variables or neurons using methods such as the Tibshirani LASSO, our new algorithm outperforms known methods.
引用
收藏
页数:33
相关论文
共 50 条
  • [1] A LOOK AT FINDING POINTS OF A CONVEX POLYHEDRON USING RELAXATION AND SUBGRADIENT PROCEDURES
    KOEHLER, GJ
    KUMAR, S
    [J]. OPERATIONS RESEARCH, 1980, 28 (02) : 424 - 430
  • [2] USING MACHINE LEARNING ALGORITHMS TO IDENTIFY OPEN ABDOMEN PROCEDURES IN ADMINISTRATIVE DATABASES
    Brat, Gabriel
    Beam, Andrew
    Salim, Ali
    Christopher, Kenneth
    [J]. CRITICAL CARE MEDICINE, 2015, 43 (12)
  • [3] Superquantiles at Work: Machine Learning Applications and Efficient Subgradient Computation
    Yassine Laguel
    Krishna Pillutla
    Jérôme Malick
    Zaid Harchaoui
    [J]. Set-Valued and Variational Analysis, 2021, 29 : 967 - 996
  • [4] Algorithms for Machine Learning
    Hsu, Daniel
    [J]. IEEE INTELLIGENT SYSTEMS, 2016, 31 (01) : 60 - 60
  • [5] Superquantiles at Work: Machine Learning Applications and Efficient Subgradient Computation
    Laguel, Yassine
    Pillutla, Krishna
    Malick, Jerome
    Harchaoui, Zaid
    [J]. SET-VALUED AND VARIATIONAL ANALYSIS, 2021, 29 (04) : 967 - 996
  • [6] Dual subgradient algorithms for large-scale nonsmooth learning problems
    Cox, Bruce
    Juditsky, Anatoli
    Nemirovski, Arkadi
    [J]. MATHEMATICAL PROGRAMMING, 2014, 148 (1-2) : 143 - 180
  • [7] Long-Term-Based Road Blackspot Screening Procedures by Machine Learning Algorithms
    Fiorentini, Nicholas
    Losa, Massimo
    [J]. SUSTAINABILITY, 2020, 12 (15)
  • [8] The promises and perils of machine learning algorithms to reduce bias and discrimination in personnel selection procedures
    Hiemstra, Annemarie M. F.
    Cassel, Tatjana
    Born, Marise Ph
    Liem, Cynthia C. S.
    [J]. GEDRAG & ORGANISATIE, 2020, 33 (04): : 279 - 299
  • [9] A RELAXATION SUBGRADIENT METHOD OF LINEARIZATION
    NASTASENKO, OA
    [J]. VESTNIK LENINGRADSKOGO UNIVERSITETA SERIYA MATEMATIKA MEKHANIKA ASTRONOMIYA, 1982, (04): : 97 - 99
  • [10] Dual subgradient algorithms for large-scale nonsmooth learning problems
    Bruce Cox
    Anatoli Juditsky
    Arkadi Nemirovski
    [J]. Mathematical Programming, 2014, 148 : 143 - 180