Proximal Quasi-Newton for Computationally Intensive l1-regularized M-estimators

被引:0
|
作者
Zhong, Kai [1 ]
Yen, Ian E. H. [2 ]
Dhillon, Inderjit S. [2 ]
Ravikumar, Pradeep [2 ]
机构
[1] Univ Texas Austin, Inst Computat Engn & Sci, Austin, TX 78712 USA
[2] Univ Texas Austin, Dept Comp Sci, Austin, TX 78712 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the class of optimization problems arising from computationally intensive l(1)-regularized M-estimators, where the function or gradient values are very expensive to compute. A particular instance of interest is the l(1)-regularized MLE for learning Conditional Random Fields (CRFs), which are a popular class of statistical models for varied structured prediction problems such as sequence labeling, alignment, and classification with label taxonomy. l(1)-regularized MLEs for CRFs are particularly expensive to optimize since computing the gradient values requires an expensive inference step. In this work, we propose the use of a carefully constructed proximal quasi-Newton algorithm for such computationally intensive M-estimation problems, where we employ an aggressive active set selection technique. In a key contribution of the paper, we show that the proximal quasi-Newton method is provably super-linearly convergent, even in the absence of strong convexity, by leveraging a restricted variant of strong convexity. In our experiments, the proposed algorithm converges considerably faster than current state-of-the-art on the problems of sequence labeling and hierarchical classification.
引用
收藏
页数:9
相关论文
共 22 条
  • [1] Sparsistency of l1-Regularized M-Estimators
    Li, Yen-Huan
    Scarlett, Jonathan
    Ravikumar, Pradeep
    Cevher, Volkan
    [J]. ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 38, 2015, 38 : 644 - 652
  • [2] IMRO: A PROXIMAL QUASI-NEWTON METHOD FOR SOLVING l1-REGULARIZED LEAST SQUARES PROBLEMS
    Karimi, Sahar
    Vavasis, Stephen
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2017, 27 (02) : 583 - 615
  • [3] An active-set proximal quasi-Newton algorithm for l1-regularized minimization over a sphere constraint
    Shen, Chungen
    Mi, Ling
    Zhang, Lei-Hong
    [J]. OPTIMIZATION, 2022, 71 (16) : 4623 - 4664
  • [4] A PROXIMAL QUASI-NEWTON TRUST-REGION METHOD FOR NONSMOOTH REGULARIZED OPTIMIZATION
    Aravkin, Aleksandr Y.
    Baraldi, Robert
    Orban, Dominique
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2022, 32 (02) : 900 - 929
  • [5] l1-Regularized full-waveform inversion with prior model information based on orthant-wise limited memory quasi-Newton method
    Dai, Meng-Xue
    Chen, Jing-Bo
    Cao, Jian
    [J]. JOURNAL OF APPLIED GEOPHYSICS, 2017, 142 : 49 - 57
  • [6] Proximal quasi-Newton methods for regularized convex optimization with linear and accelerated sublinear convergence rates
    Hiva Ghanbari
    Katya Scheinberg
    [J]. Computational Optimization and Applications, 2018, 69 : 597 - 627
  • [7] Proximal quasi-Newton methods for regularized convex optimization with linear and accelerated sublinear convergence rates
    Ghanbari, Hiva
    Scheinberg, Katya
    [J]. COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2018, 69 (03) : 597 - 627
  • [8] Convergence of Common Proximal Methods for l1-Regularized Least Squares
    Tao, Shaozhe
    Boley, Daniel
    Zhang, Shuzhong
    [J]. PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 3849 - 3855
  • [9] Orthant Based Proximal Stochastic Gradient Method for l1-Regularized Optimization
    Chen, Tianyi
    Ding, Tianyu
    Ji, Bo
    Wang, Guanyi
    Shi, Yixin
    Tian, Jing
    Yi, Sheng
    Tu, Xiao
    Zhu, Zhihui
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2020, PT III, 2021, 12459 : 57 - 73
  • [10] CONVERGENCE OF AN INERTIAL PROXIMAL METHOD FOR L1-REGULARIZED LEAST-SQUARES
    Johnstone, Patrick R.
    Moulin, Pierre
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 3566 - 3570