Unsupervised Tree Boosting for Learning Probability Distributions

被引:0
|
作者
Awaya, Naoki [1 ]
Ma, Li [2 ]
机构
[1] Waseda Univ, Sch Polit Sci & Econ, Shinjuku City, Tokyo 1698050, Japan
[2] Duke Univ, Dept Stat Sci, Durham, NC 27708 USA
关键词
generative models; normalizing flows; additive models; density estimation; ensemble methods; recursive partitioning; POLYA TREE;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We propose an unsupervised tree boosting algorithm for inferring the underlying sampling distribution of an i.i.d. sample based on fitting additive tree ensembles in a manner analogous to supervised tree boosting. Integral to the algorithm is a new notion of "addition" on probability distributions that leads to a coherent notion of "residualization", i.e., subtracting a probability distribution from an observation to remove the distributional structure from the sampling distribution of the latter. We show that these notions arise naturally for univariate distributions through cumulative distribution function (CDF) transforms and compositions due to several "group-like" properties of univariate CDFs. While the traditional multivariate CDF does not preserve these properties, a new definition of multivariate CDF can restore these properties, thereby allowing the notions of "addition" and "residualization" to be formulated for multivariate settings as well. This then gives rise to the unsupervised boosting algorithm based on forward-stagewise fitting of an additive tree ensemble, which sequentially reduces the Kullback-Leibler divergence from the truth. The algorithm allows analytic evaluation of the fitted density and outputs a generative model that can be readily sampled from. We enhance the algorithm with scale-dependent shrinkage and a two-stage strategy that separately fits the marginals and the copula. The algorithm then performs competitively with state-of-the-art deep-learning approaches in multivariate density estimation on multiple benchmark data sets.
引用
收藏
页数:52
相关论文
共 50 条
  • [31] Unsupervised learning via mixtures of skewed distributions with hypercube contours
    Franczak, Brian C.
    Tortora, Cristina
    Browne, Ryan P.
    McNicholas, Paul D.
    [J]. PATTERN RECOGNITION LETTERS, 2015, 58 : 69 - 76
  • [33] Probability Distributions Achieving the Equilibrium of an AND-OR Tree under Directional Algorithms
    Suzuki, Toshio
    Nakamura, Ryota
    [J]. INTERNATIONAL MULTICONFERENCE OF ENGINEERS AND COMPUTER SCIENTISTS, IMECS 2012, VOL I, 2012, : 194 - 199
  • [34] The Estimation of Tree Posterior Probabilities Using Conditional Clade Probability Distributions
    Larget, Bret
    [J]. SYSTEMATIC BIOLOGY, 2013, 62 (04) : 501 - 511
  • [35] How effective is incidental learning of the shape of probability distributions?
    Tran, Randy
    Vul, Edward
    Pashler, Harold
    [J]. ROYAL SOCIETY OPEN SCIENCE, 2017, 4 (08):
  • [36] Equivalences between learning of data and probability distributions, and their applications
    Barmpalias, George
    Fang, Nan
    Stephan, Frank
    [J]. INFORMATION AND COMPUTATION, 2018, 262 : 123 - 140
  • [37] Learning non-stationary conditional probability distributions
    Husmeier, D
    [J]. NEURAL NETWORKS, 2000, 13 (03) : 287 - 290
  • [38] PAC learning of probability distributions over a discrete domain
    Magnoni, L
    Mirolli, M
    Montagna, F
    Simi, G
    [J]. THEORETICAL COMPUTER SCIENCE, 2003, 299 (1-3) : 37 - 63
  • [39] An Elastic Gradient Boosting Decision Tree for Concept Drift Learning
    Wang, Kun
    Liu, Anjin
    Lu, Jie
    Zhang, Guangquan
    Xiong, Li
    [J]. AI 2020: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 12576 : 420 - 432
  • [40] A BFS-Tree of ranking references for unsupervised manifold learning
    Guimaraes Pedronette, Daniel Carlos
    Valem, Lucas Pascotti
    Torres, Ricardo da S.
    [J]. PATTERN RECOGNITION, 2021, 111