Semivarying coefficient least-squares support vector regression for analyzing high-dimensional gene-environmental data

被引:2
|
作者
Shim, Jooyong [1 ]
Hwang, Changha [2 ]
Jeong, Sunjoo [3 ]
Sohn, Insuk [4 ]
机构
[1] Inje Univ, Inst Stat Informat, Dept Stat, Kyungnam, South Korea
[2] Dankook Univ, Dept Appl Stat, Yongin, Gyeonggido, South Korea
[3] Dankook Univ, Dept Bioconvergent Sci & Technol, Yongin, Gyeonggido, South Korea
[4] Samsung Med Ctr, Stat & Data Ctr, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
Generalized cross validation; gene-environment interaction; least-squares support vector regression; main effect; semiparametric regression; semivarying coefficient; survival data; variable selection; varying coefficient regression; CENSORED-DATA; SNP DATA; MODELS;
D O I
10.1080/02664763.2017.1371676
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In the context of genetics and genomic medicine, gene-environment (GxE) interactions have a great impact on the risk of human diseases. Some existing methods for identifying GxE interactions are considered to be limited, since they analyze one or a few number of G factors at a time, assume linear effects of E factors, and use inefficient selection methods. In this paper, we propose a new method to identify significant main effects and GxE interactions. This is based on a semivarying coefficient least-squares support vector regression (LS-SVR) technique, which is devised by utilizing flexible semiparametric LS-SVR approach for censored survival data. This semivarying coefficient model is used to deal with the nonlinear effects of E factors. We also derive a generalized cross validation (GCV) function for determining the optimal values of hyperparameters of the proposed method. This GCV function is also used to identify significant main effects and GxE interactions. The proposed method is evaluated through numerical studies.
引用
收藏
页码:1370 / 1381
页数:12
相关论文
共 50 条
  • [1] Thresholding least-squares inference in high-dimensional regression models
    Giurcanu, Mihai
    ELECTRONIC JOURNAL OF STATISTICS, 2016, 10 (02): : 2124 - 2156
  • [2] SPARSE LEAST TRIMMED SQUARES REGRESSION FOR ANALYZING HIGH-DIMENSIONAL LARGE DATA SETS
    Alfons, Andreas
    Croux, Christophe
    Gelper, Sarah
    ANNALS OF APPLIED STATISTICS, 2013, 7 (01): : 226 - 248
  • [3] Nonnegative Least-Squares Methods for the Classification of High-Dimensional Biological Data
    Li, Yifeng
    Ngom, Alioune
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2013, 10 (02) : 447 - 456
  • [4] Incremental algorithm for regression least-squares support vector machines
    Zhang, Xuefeng
    Yuan, Yubo
    Huang, Tingzhu
    Advances in Matrix Theory and Applications, 2006, : 396 - 399
  • [5] NONLINEAR UNMIXING OF HYPERSPECTRAL DATA WITH PARTIALLY LINEAR LEAST-SQUARES SUPPORT VECTOR REGRESSION
    Chen, Jie
    Richard, Cedric
    Ferrari, Andre
    Honeine, Paul
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 2174 - 2178
  • [6] High-Dimensional Least-Squares with Perfect Positive Correlation
    Huang, Zhiyong
    Luo, Ziyan
    Xiu, Naihua
    ASIA-PACIFIC JOURNAL OF OPERATIONAL RESEARCH, 2019, 36 (04)
  • [7] Penalized least-squares estimation for regression coefficients in high-dimensional partially linear models
    Ni, Huey-Fan
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2012, 142 (02) : 379 - 389
  • [8] PARTIAL LEAST SQUARES PREDICTION IN HIGH-DIMENSIONAL REGRESSION
    Cook, R. Dennis
    Forzani, Liliana
    ANNALS OF STATISTICS, 2019, 47 (02): : 884 - 908
  • [9] Multi-output least-squares support vector regression machines
    Xu, Shuo
    An, Xin
    Qiao, Xiaodong
    Zhu, Lijun
    Li, Lin
    PATTERN RECOGNITION LETTERS, 2013, 34 (09) : 1078 - 1084
  • [10] Varying coefficient modeling via least squares support vector regression
    Shim, Jooyong
    Hwang, Changha
    NEUROCOMPUTING, 2015, 161 : 254 - 259