Minimax-optimal classification with dyadic decision trees

被引:59
|
作者
Scott, C [1 ]
Nowak, RD
机构
[1] Rice Univ, Dept Stat, Houston, TX 77005 USA
[2] Univ Wisconsin, Dept Elect & Comp Engn, Madison, WI 53706 USA
基金
美国国家科学基金会;
关键词
complexity regularization; decision trees; feature rejection; generalization error bounds; manifold learning; minimax optimality; pruning; rates of convergence; recursive dyadic partitions; statistical learning theory;
D O I
10.1109/TIT.2006.871056
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Decision trees are among the most popular types of classifiers, with interpretability and ease of implementation being among their chief attributes. Despite the widespread use of decision trees, theoretical analysis of their performance has only begun to emerge in recent years. In this paper, it is shown that a new family of decision trees, dyadic decision trees (DDTs), attain nearly optimal (in a minimax sense) rates of convergence for a broad range of classification problems. Furthermore, DDTs are surprisingly adaptive in three important respects: they automatically 1) adapt to favorable conditions near the Bayes decision boundary; 2) focus on data distributed on lower dimensional manifolds; and 3) reject irrelevant features. DDTs are constructed by penalized empirical risk minimization using a new data-dependent penalty and may be computed exactly with computational complexity that is nearly linear in the training sample size. DDTs comprise the first classifiers known to achieve nearly optimal rates for the diverse class of distributions studied here while also being practical and implementable. This is also the first study (of which we are aware) to consider rates for adaptation to intrinsic data dimension and relevant features.
引用
收藏
页码:1335 / 1353
页数:19
相关论文
共 50 条
  • [1] Near-minimax optimal classification with dyadic classification trees
    Scott, C
    Nowak, R
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 16, 2004, 16 : 1117 - 1124
  • [2] Optimal dyadic decision trees
    Blanchard, G.
    Schaefer, C.
    Rozenholc, Y.
    Mueller, K. -R.
    [J]. MACHINE LEARNING, 2007, 66 (2-3) : 209 - 241
  • [3] Optimal dyadic decision trees
    G. Blanchard
    C. Schäfer
    Y. Rozenholc
    K.-R. Müller
    [J]. Machine Learning, 2007, 66 : 209 - 241
  • [4] Minimax-Optimal Location Estimation
    Gupta, Shivam
    Lee, Jasper C. H.
    Price, Eric
    Valiant, Paul
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Algorithms for optimal dyadic decision trees
    Hush, Don
    Porter, Reid
    [J]. MACHINE LEARNING, 2010, 80 (01) : 85 - 107
  • [6] Algorithms for optimal dyadic decision trees
    Don Hush
    Reid Porter
    [J]. Machine Learning, 2010, 80 : 85 - 107
  • [7] Analysis and design of minimax-optimal interpolators
    Choi, H
    Munson, DC
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 1998, 46 (06) : 1571 - 1579
  • [8] Minimax-optimal Inference from Partial Rankings
    Hajek, Bruce
    Oh, Sewoong
    Xu, Jiaming
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [9] Nearly Minimax-Optimal Regret for Linearly Parameterized Bandits
    Li, Yingkai
    Wang, Yining
    Zhou, Yuan
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2024, 70 (01) : 372 - 388
  • [10] MINIMAX-OPTIMAL NONPARAMETRIC REGRESSION IN HIGH DIMENSIONS
    Yang, Yun
    Tokdar, Surya T.
    [J]. ANNALS OF STATISTICS, 2015, 43 (02): : 652 - 674