Hybrid Dirichlet mixture models for functional data

被引:61
|
作者
Petrone, Sonia [1 ]
Guindani, Michele [2 ]
Gelfand, Alan E. [3 ]
机构
[1] Univ Bocconi, Ist Metodi Quantitat, I-20135 Milan, Italy
[2] Univ New Mexico, Albuquerque, NM 87131 USA
[3] Duke Univ, Durham, NC 27706 USA
基金
美国国家科学基金会;
关键词
Bayesian non-parametrics; Dependent random partitions; Dirichlet process; Finite mixture models; Gaussian process; Labelling measures; Species sampling priors;
D O I
10.1111/j.1467-9868.2009.00708.x
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In functional data analysis, curves or surfaces are observed, up to measurement error, at a finite set of locations, for, say, a sample of n individuals. Often, the curves are homogeneous, except perhaps for individual-specific regions that provide heterogeneous behaviour (e.g. 'damaged' areas of irregular shape on an otherwise smooth surface). Motivated by applications with functional data of this nature, we propose a Bayesian mixture model, with the aim of dimension reduction, by representing the sample of n curves through a smaller set of canonical curves. We propose a novel prior on the space of probability measures for a random curve which extends the popular Dirichlet priors by allowing local clustering: non-homogeneous portions of a curve can be allocated to different clusters and the n individual curves can be represented as recombinations (hybrids) of a few canonical curves. More precisely, the prior proposed envisions a conceptual hidden factor with k-levels that acts locally on each curve. We discuss several models incorporating this prior and illustrate its performance with simulated and real data sets. We examine theoretical properties of the proposed finite hybrid Dirichlet mixtures, specifically, their behaviour as the number of the mixture components goes to infinity and their connection with Dirichlet process mixtures.
引用
收藏
页码:755 / 782
页数:28
相关论文
共 50 条
  • [1] Dirichlet process mixture models for insurance loss data
    Hong, Liang
    Martin, Ryan
    SCANDINAVIAN ACTUARIAL JOURNAL, 2018, (06) : 545 - 554
  • [2] Unsupervised learning of Dirichlet process mixture models with missing data
    Xunan ZHANG
    Shiji SONG
    Lei ZHU
    Keyou YOU
    Cheng WU
    Science China(Information Sciences), 2016, 59 (01) : 161 - 174
  • [3] Sampling in Dirichlet Process Mixture Models for Clustering Streaming Data
    Dinari, Or
    Freifeld, Oren
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151 : 818 - 835
  • [4] An optimal data ordering scheme for Dirichlet process mixture models
    Wang, Xue
    Walker, Stephen G.
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2017, 112 : 42 - 52
  • [5] CLASSIFICATION OF MULTIVARIATE DATA USING DIRICHLET PROCESS MIXTURE MODELS
    Djuric, Petar M.
    Ferrari, Andre
    2012 CONFERENCE RECORD OF THE FORTY SIXTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS AND COMPUTERS (ASILOMAR), 2012, : 441 - 445
  • [6] Unsupervised learning of Dirichlet process mixture models with missing data
    Zhang, Xunan
    Song, Shiji
    Zhu, Lei
    You, Keyou
    Wu, Cheng
    SCIENCE CHINA-INFORMATION SCIENCES, 2016, 59 (01) : 1 - 14
  • [7] Scalable Estimation of Dirichlet Process Mixture Models on Distributed Data
    Wang, Ruohui
    Lin, Dahua
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 4632 - 4639
  • [8] Dirichlet Process Mixture Models with Pairwise Constraints for Data Clustering
    Li C.
    Rana S.
    Phung D.
    Venkatesh S.
    Annals of Data Science, 2016, 3 (2) : 205 - 223
  • [9] DIRICHLET PROCESS MIXTURE MODELS FOR CLUSTERING I-VECTOR DATA
    Seshadri, Shreyas
    Remes, Ulpu
    Rasanen, Okko
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 5470 - 5474
  • [10] Dirichlet process mixture models for non-stationary data streams
    Casado, Ioar
    Perez, Aritz
    2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2022, : 873 - 878