LEAST QUANTILE REGRESSION VIA MODERN OPTIMIZATION

被引:30
|
作者
Bertsimas, Dimitris [1 ,2 ]
Mazumder, Rahul [3 ]
机构
[1] MIT, MIT Sloan Sch Management, Cambridge, MA 02139 USA
[2] MIT, Operat Res Ctr, Cambridge, MA 02139 USA
[3] Columbia Univ, Dept Stat, New York, NY 10027 USA
来源
ANNALS OF STATISTICS | 2014年 / 42卷 / 06期
关键词
Least median of squares; robust statistics; least quantile regression; algorithms; mixed integer programming; global optimization; continuous optimization; ROBUST REGRESSION; APPROXIMATION ALGORITHM; SQUARES ESTIMATE;
D O I
10.1214/14-AOS1223
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We address the Least Quantile of Squares (LQS) (and in particular the Least Median of Squares) regression problem using modern optimization methods. We propose a Mixed Integer Optimization (MIO) formulation of the LQS problem which allows us to find a provably global optimal solution for the LQS problem. Our MIO framework has the appealing characteristic that if we terminate the algorithm early, we obtain a solution with a guarantee on its sub-optimality. We also propose continuous optimization methods based on first-order subdifferential methods, sequential linear optimization and hybrid combinations of them to obtain near optimal solutions to the LQS problem. The MIO algorithm is found to benefit significantly from high quality solutions delivered by our continuous optimization based methods. We further show that the MIO approach leads to (a) an optimal solution for any dataset, where the data-points (y(i), x(i))'s are not necessarily in general position, (b) a simple proof of the breakdown point of the LQS objective value that holds for any dataset and (c) an extension to situations where there are polyhedral constraints on the regression coefficient vector. We report computational results with both synthetic and real-world datasets showing that the MIO algorithm with warm starts from the continuous optimization methods solve small (n = 100) and medium (n = 500) size problems to provable optimality in under two hours, and outperform all publicly available methods for large-scale (n = 10,000) LQS problems.
引用
收藏
页码:2494 / 2525
页数:32
相关论文
共 50 条
  • [1] Quantile regression via iterative least squares computations
    Muggeo, Vito M. R.
    Sciandra, Mariangela
    Augugliaro, Luigi
    [J]. JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2012, 82 (11) : 1557 - 1569
  • [2] The least trimmed quantile regression
    Neykov, N. M.
    Cizek, P.
    Filzmoser, P.
    Neytchev, P. N.
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2012, 56 (06) : 1757 - 1770
  • [3] Unbiasedness in least quantile regression
    Tasche, D
    [J]. DEVELOPMENTS IN ROBUST STATISTICS, 2003, : 377 - 386
  • [4] Regression shrinkage and selection via least quantile shrinkage and selection operator
    Daneshvar, Alireza
    Mousa, Golalizadeh
    [J]. PLOS ONE, 2023, 18 (02):
  • [5] Distributed quantile regression in decentralized optimization
    Shen, Lin
    Chao, Yue
    Ma, Xuejun
    [J]. INFORMATION SCIENCES, 2023, 643
  • [6] Construction of optimal designs for quantile regression model via particle swarm optimization
    Yi Zhai
    Chen Xing
    Zhide Fang
    [J]. Journal of the Korean Statistical Society, 2023, 52 : 921 - 943
  • [7] Construction of optimal designs for quantile regression model via particle swarm optimization
    Zhai, Yi
    Xing, Chen
    Fang, Zhide
    [J]. JOURNAL OF THE KOREAN STATISTICAL SOCIETY, 2023, 52 (04) : 921 - 943
  • [8] Quantile Regression via the EM Algorithm
    Zhou, Ying-Hui
    Ni, Zhong-Xin
    Li, Yong
    [J]. COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2014, 43 (10) : 2162 - 2172
  • [9] Conformalized Fairness via Quantile Regression
    Liu, Meichen
    Ding, Lei
    Yu, Dengdeng
    Liu, Wulong
    Kong, Linglong
    Jiang, Bei
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [10] Quantile regression via an MM algorithm
    Hunter, DR
    Lange, K
    [J]. JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2000, 9 (01) : 60 - 77