Functional sufficient dimension reduction through information maximization with application to classification

被引:0
|
作者
Li, Xinyu [1 ]
Xu, Jianjun [2 ,4 ]
Cheng, Haoyang [3 ]
机构
[1] Univ Sci & Technol China, Int Inst Finance, Sch Management, Hefei, Anhui, Peoples R China
[2] Hefei Univ Technol, Sch Math, Hefei, Anhui, Peoples R China
[3] Quzhou Univ, Coll Elect & Informat Engn, Quzhou, Zhejiang, Peoples R China
[4] Hefei Univ Technol, Sch Math, Hefei 230601, Anhui, Peoples R China
基金
中国博士后科学基金;
关键词
Functional data classification; functional sufficient dimension reduction; mutual information; square loss mutual information; density ratio; REGRESSION; RECOGNITION; DIVERGENCE; PREDICTION; MORTALITY; COHORT; RATES;
D O I
10.1080/02664763.2024.2335570
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Considering the case where the response variable is a categorical variable and the predictor is a random function, two novel functional sufficient dimensional reduction (FSDR) methods are proposed based on mutual information and square loss mutual information. Compared to the classical FSDR methods, such as functional sliced inverse regression and functional sliced average variance estimation, the proposed methods are appealing because they are capable of estimating multiple effective dimension reduction directions in the case of a relatively small number of categories, especially for the binary response. Moreover, the proposed methods do not require the restrictive linear conditional mean assumption and the constant covariance assumption. They avoid the inverse problem of the covariance operator which is often encountered in the functional sufficient dimension reduction. The functional principal component analysis with truncation be used as a regularization mechanism. Under some mild conditions, the statistical consistency of the proposed methods is established. Simulation studies and real data analyzes are used to evaluate the finite sample properties of our methods.
引用
收藏
页码:3059 / 3101
页数:43
相关论文
共 50 条
  • [21] On sufficient dimension reduction with missing responses through estimating equations
    Dong, Yuexiao
    Xia, Qi
    Tang, Cheng Yong
    Li, Zeda
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2018, 126 : 67 - 77
  • [22] Online sufficient dimension reduction through sliced inverse regression
    Cai, Zhanrui
    Li, Runze
    Zhu, Liping
    Journal of Machine Learning Research, 2020, 21
  • [23] Online Sufficient Dimension Reduction Through Sliced Inverse Regression
    Cai, Zhanrui
    Li, Runze
    Zhu, Liping
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [24] Sufficient dimension reduction and prediction through cumulative slicing PFC
    Xu, Xinyi
    Li, Xiangjie
    Zhang, Jingxiao
    JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2018, 88 (06) : 1172 - 1190
  • [25] Sufficient dimension reduction in regressions through cumulative Hessian directions
    Zhang, Li-Mei
    Zhu, Li-Ping
    Zhu, Li-Xing
    STATISTICS AND COMPUTING, 2011, 21 (03) : 325 - 334
  • [26] Sufficient dimension reduction in regressions through cumulative Hessian directions
    Li-Mei Zhang
    Li-Ping Zhu
    Li-Xing Zhu
    Statistics and Computing, 2011, 21 : 325 - 334
  • [27] Sufficient dimension reduction through discretization-expectation estimation
    Zhu, Liping
    Wang, Tao
    Zhu, Lixing
    Ferre, Louis
    BIOMETRIKA, 2010, 97 (02) : 295 - 304
  • [28] A brief review of linear sufficient dimension reduction through optimization
    Dong, Yuexiao
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2021, 211 : 154 - 161
  • [29] Principal weighted logistic regression for sufficient dimension reduction in binary classification
    Boyoung Kim
    Seung Jun Shin
    Journal of the Korean Statistical Society, 2019, 48 : 194 - 206
  • [30] Sufficient dimension reduction for classification using principal optimal transport direction
    Meng, Cheng
    Yu, Jun
    Zhang, Jingyi
    Ma, Ping
    Zhong, Wenxuan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33