Efficient distribution-free population learning of simple concepts

被引:1
|
作者
Nakamura, A [1 ]
Takeuchi, J [1 ]
Abe, N [1 ]
机构
[1] NEC Corp Ltd, C&C Media Res Labs, RWCP, Theory NEC Lab,Miyamae Ku, Kawasaki, Kanagawa 2168555, Japan
关键词
Majority Vote; Concept Class; Vote Scheme; Target Concept; Hypothesis Class;
D O I
10.1023/A:1018908122958
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider a variant of the 'population learning model' proposed by Kearns and Seung [8], in which the learner is required to be 'distribution-free' as well as computationally efficient. A population learner receives as input hypotheses from a large population of agents and produces as output its final hypothesis. Each agent is assumed to independently obtain labeled sample for the target concept and output a hypothesis. A polynomial time population learner is said to PAC-learn a concept class, if its hypothesis is probably approximately correct whenever the population size exceeds a certain bound which is polynomial, even if the sample size for each agent is fixed at some constant. We exhibit some general population learning strategies, and some simple concept classes that can be learned by them. These strategies include the 'supremum hypothesis finder', the 'minimum superset finder' (a special case of the 'supremum hypothesis finder'), and various voting schemes. When coupled with appropriate agent algorithms, these strategies can learn a variety of simple concept classes, such as the 'high-low game', conjunctions, axis-parallel rectangles and others. We give upper bounds on the required population size for each of these cases, and show that these systems can be used to obtain a speed up from the ordinary PAC-learning model [11], with appropriate choices of sample and population sizes. With the population learner restricted to be a voting scheme, what we have is effectively a model of 'population prediction', in which the learner is to predict the value of the target concept at an arbitrarily drawn point, as a threshold function of the predictions made by its agents on the same point. We show that the population learning model is strictly more powerful than the population prediction model. Finally, we consider a variant of this model with classification noise, and exhibit a population learner for the class of conjunctions in this model.
引用
收藏
页码:53 / 82
页数:30
相关论文
共 50 条
  • [41] Distribution-Free Junta Testing
    Liu, Zhengyang
    Chen, Xi
    Servedio, Rocco A.
    Sheng, Ying
    Xie, Jinyu
    STOC'18: PROCEEDINGS OF THE 50TH ANNUAL ACM SIGACT SYMPOSIUM ON THEORY OF COMPUTING, 2018, : 749 - 759
  • [42] DISTRIBUTION-FREE TESTS OF RANDOMNESS
    BELL, CB
    DONOGHUE, JF
    SANKHYA-THE INDIAN JOURNAL OF STATISTICS SERIES A, 1969, 31 (JUN): : 157 - 176
  • [43] DISTRIBUTION-FREE MULTIPLE COMPARISONS
    NEMENYI, P
    BIOMETRICS, 1962, 18 (02) : 263 - &
  • [44] DISTRIBUTION-FREE CONFIDENCE INTERVALS
    NOETHER, GE
    AMERICAN STATISTICIAN, 1972, 26 (01): : 39 - &
  • [45] ON THE STRUCTURE OF DISTRIBUTION-FREE STATISTICS
    BELL, CB
    ANNALS OF MATHEMATICAL STATISTICS, 1960, 31 (03): : 703 - 709
  • [46] Distribution-Free Prediction Sets
    Lei, Jing
    Robins, James
    Wasserman, Larry
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2013, 108 (501) : 278 - 287
  • [47] DISTRIBUTION-FREE CONFIDENCE INTERVALS
    NOETHER, GE
    BIOMETRICS, 1972, 28 (04) : 1178 - 1178
  • [48] On a distribution-free quantile estimator
    Huang, ML
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2001, 37 (04) : 477 - 486
  • [49] Distribution-free option pricing
    De Schepper, Ann
    Heijnen, Bart
    INSURANCE MATHEMATICS & ECONOMICS, 2007, 40 (02): : 179 - 199
  • [50] DISTRIBUTION-FREE TOLERANCE LIMITS
    GOVINDARAJULU, Z
    NAVAL RESEARCH LOGISTICS, 1977, 24 (02) : 381 - 384