Online and Distribution-Free Robustness: Regression and Contextual Bandits with Huber Contamination

被引:9
|
作者
Chen, Sitan [1 ]
Koehler, Frederic [1 ]
Moitra, Ankur [2 ]
Yau, Morris [2 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
[2] MIT, 77 Massachusetts Ave, Cambridge, MA 02139 USA
关键词
robust statistics; regression; contextual bandits; online learning; Huber contamination; ASYMPTOTICS;
D O I
10.1109/FOCS52979.2021.00072
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In this work we revisit two classic high-dimensional online learning problems, namely linear regression and contextual bandits, from the perspective of adversarial robustness. Existing works in algorithmic robust statistics make strong distributional assumptions that ensure that the input data is evenly spread out or comes from a nice generative model. Is it possible to achieve strong robustness guarantees even without distributional assumptions altogether, where the sequence of tasks we are asked to solve is adaptively and adversarially chosen? We answer this question in the affirmative for both linear regression and contextual bandits. In fact our algorithms succeed where conventional methods fail. In particular we show strong lower bounds against Huber regression and more generally any convex M-estimator. Our approach is based on a novel alternating minimization scheme that interleaves ordinary least-squares with a simple convex program that finds the optimal reweighting of the distribution under a spectral constraint. Our results obtain essentially optimal dependence on the contamination level eta, reach the optimal breakdown point, and naturally apply to infinite dimensional settings where the feature vectors are represented implicitly via a kernel map.
引用
收藏
页码:684 / 695
页数:12
相关论文
共 50 条