Learning General Halfspaces with General Massart Noise under the Gaussian Distribution

被引:7
|
作者
Diakonikolas, Ilias [1 ]
Kane, Daniel M. [2 ]
Kontonis, Vasilis [1 ]
Tzamos, Christos [1 ]
Zarifis, Nikos [1 ]
机构
[1] Univ Wisconsin, Madison, WI 53706 USA
[2] Univ Calif San Diego, San Diego, CA 92103 USA
关键词
PAC learning; Halfspaces; Massart Noise; MODEL;
D O I
10.1145/3519935.3519970
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
We study the problem of PAC learning halfspaces on R-d with Massart noise under the Gaussian distribution. In the Massart model, an adversary is allowed to flip the label of each point x with unknown probability eta(x) <= eta, for some parameter eta is an element of[0,1/2]. The goal is to find a hypothesis with misclassification error of OPT+epsilon, where OPT is the error of the target halfspace. This problem had been previously studied under two assumptions: (i) the target halfspace is homogeneous (i.e., the separating hyperplane goes through the origin), and (ii) the parameter eta is strictly smaller than 1/2. Prior to this work, no nontrivial bounds were known when either of these assumptions is removed. We study the general problem and establish the following: For eta < 1/2, we give a learning algorithm for general halfspaces with sample and computational complexity d(O eta(log(1/gamma)))poly(1/epsilon), where gamma :=max{epsilon,min{Pr[f(x)=1],Pr[f(x)=-1]}} is the bias of the target halfspace f. Prior efficient algorithms could only handle the special case of gamma=1/2. Interestingly, we establish a qualitatively matching lower bound of d(Omega(log(1/gamma))) on the complexity of any Statistical Query (SQ) algorithm. For eta=1/2, we give a learning algorithm for general halfspaces with sample and computational complexity O-epsilon(1)d(O(log(1/epsilon))). This result is new even for the subclass of homogeneous halfspaces; prior algorithms for homogeneous Massart halfspaces provide vacuous guarantees for eta=1/2. We complement our upper bound with a nearly-matching SQ lower bound of d(Omega(log(1/epsilon))), which holds even for the special case of homogeneous halfspaces. Taken together, our results qualitatively characterize the complexity of learning general halfspaces with general Massart noise under Gaussian marginals. Our techniques rely on determining the existence (or non-existence) of low-degree polynomials whose expectations distinguish Massart halfspaces from random noise.
引用
收藏
页码:874 / 885
页数:12
相关论文
共 50 条
  • [1] Distribution-Independent PAC Learning of Halfspaces with Massart Noise
    Diakonikolas, Ilias
    Gouleakis, Themis
    Tzamos, Christos
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Cryptographic Hardness of Learning Halfspaces with Massart Noise
    Diakonikolas, Ilias
    Kane, Daniel M.
    Manurangsi, Pasin
    Ren, Lisheng
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [3] Learning General Halfspaces with Adversarial Label Noise via Online Gradient Descent
    Diakonikolas, Ilias
    Kontonis, Vasilis
    Tzamos, Christos
    Zarifis, Nikos
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [4] Learning Halfspaces with Malicious Noise
    Klivans, Adam R.
    Long, Philip M.
    Servedio, Rocco A.
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2009, 10 : 2715 - 2740
  • [5] Near-Optimal Bounds for Learning Gaussian Halfspaces with Random Classification Noise
    Diakonikolas, Ilias
    Diakonikolas, Jelena
    Kane, Daniel M.
    Wang, Puqian
    Zarifis, Nikos
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] Learning Halfspaces with Malicious Noise
    Klivans, Adam R.
    Long, Philip M.
    Servedio, Rocco A.
    [J]. AUTOMATA, LANGUAGES AND PROGRAMMING, PT I, 2009, 5555 : 609 - +
  • [7] Random oscillator with general Gaussian noise
    Mallick, K.
    [J]. PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2007, 384 (01) : 64 - 68
  • [8] Hardness of learning halfspaces with noise
    Guruswami, Venkatesan
    Raghavendra, Prasad
    [J]. 47TH ANNUAL IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, PROCEEDINGS, 2006, : 543 - +
  • [9] HARDNESS OF LEARNING HALFSPACES WITH NOISE
    Guruswami, Venkatesan
    Raghavendra, Prasad
    [J]. SIAM JOURNAL ON COMPUTING, 2009, 39 (02) : 742 - 765
  • [10] Forster Decomposition and Learning Halfspaces with Noise
    Diakonikolas, Ilias
    Kane, Daniel M.
    Tzamos, Christos
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34