SELECTIVE INFERENCE FOR SPARSE MULTITASK REGRESSION WITH APPLICATIONS IN NEUROIMAGING

被引:0
|
作者
Panigrahi, Snigdha [1 ]
Stewart, Natasha [1 ]
Sripada, Chandra [2 ]
Levina, Elizaveta [1 ]
机构
[1] Univ Michigan, Dept Stat, Ann Arbor, MI 48109 USA
[2] Univ Michigan, Dept Philosophy, Ann Arbor, MI USA
来源
ANNALS OF APPLIED STATISTICS | 2024年 / 18卷 / 01期
关键词
Multitask learning; multilevel lasso; joint sparsity; postselection inference; selective inference; neuroimaging; fMRI data; COGNITIVE CONTROL NETWORK; CEREBELLUM; MOVEMENT; DIFFERENTIATION; INTELLIGENCE; FEATURES; PREDICT;
D O I
10.1214/23-AOAS1796
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Multitask learning is frequently used to model a set of related response variables from the same set of features, improving predictive performance and modeling accuracy relative to methods that handle each response variable separately. Despite the potential of multitask learning to yield more powerful inference than single-task alternatives, prior work in this area has largely omitted uncertainty quantification. Our focus in this paper is a common multitask problem in neuroimaging, where the goal is to understand the relationship between multiple cognitive task scores (or other subject-level assessments) and brain connectome data collected from imaging. We propose a framework for selective inference to address this problem, with the flexibility to: (i) jointly identify the relevant predictors for each task through a sparsity-inducing penalty and (ii) conduct valid inference in a model based on the estimated sparsity structure. Our framework offers a new conditional procedure for inference, based on a refinement of the selection event that yields a tractable selection-adjusted likelihood. This gives an approximate system of estimating equations for maximum likelihood inference, solvable via a single convex optimization problem, and enables us to efficiently form confidence intervals with approximately the correct coverage. Applied to both simulated data and data from the Adolescent Brain Cognitive Development vals than commonly used alternatives, such as data splitting. We also demonstrate through simulations that multitask learning with selective inference can more accurately recover true signals than single-task methods.
引用
收藏
页码:445 / 467
页数:23
相关论文
共 50 条
  • [41] On Merging MobileNets for Efficient Multitask Inference
    Wu, Cheng-En
    Chan, Yi-Ming
    Chen, Chu-Song
    2019 2ND WORKSHOP ON ENERGY EFFICIENT MACHINE LEARNING AND COGNITIVE COMPUTING FOR EMBEDDED APPLICATIONS (EMC2 2019), 2019, : 16 - 20
  • [42] Multitask Learning for Sparse Failure Prediction
    Luo, Simon
    Chu, Victor W.
    Li, Zhidong
    Wang, Yang
    Zhou, Jianlong
    Chen, Fang
    Wong, Raymond K.
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2019, PT I, 2019, 11439 : 3 - 14
  • [43] Consensus inference in neuroimaging
    Hansen, LK
    Nielsen, FA
    Strother, SC
    Lange, N
    NEUROIMAGE, 2001, 13 (06) : 1212 - 1218
  • [44] Online Sparse Gaussian Process Regression and Its Applications
    Ranganathan, Ananth
    Yang, Ming-Hsuan
    Ho, Jeffrey
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2011, 20 (02) : 391 - 404
  • [45] Sparse Functional Linear Regression with Applications to Personalized Medicine
    McKeague, Ian W.
    Qian, Min
    RECENT ADVANCES IN FUNCTIONAL DATA ANALYSIS AND RELATED TOPICS, 2011, : 213 - 218
  • [46] Logistic Regression for Fuzzy Covariates: Modeling, Inference, and Applications
    Fatemeh Salmani
    S. Mahmoud Taheri
    Jin Hee Yoon
    Alireza Abadi
    Hamid Alavi Majd
    Abbas Abbaszadeh
    International Journal of Fuzzy Systems, 2017, 19 : 1635 - 1644
  • [47] Logistic Regression for Fuzzy Covariates: Modeling, Inference, and Applications
    Salmani, Fatemeh
    Taheri, S. Mahmoud
    Yoon, Jin Hee
    Abadi, Alireza
    Majd, Hamid Alavi
    Abbaszadeh, Abbas
    INTERNATIONAL JOURNAL OF FUZZY SYSTEMS, 2017, 19 (05) : 1635 - 1644
  • [48] Distributed Variational Inference in Sparse Gaussian Process Regression and Latent Variable Models
    Gal, Yarin
    van der Wilk, Mark
    Rasmussen, Carl E.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [49] Approximate Post-Selective Inference for Regression with the Group LASSO
    Panigrahi, Snigdha
    MacDonald, Peter W.
    Kessler, Daniel
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [50] Novel applications of multitask learning and multiple output regression to multiple genetic trait prediction
    He, Dan
    Kuhn, David
    Parida, Laxmi
    BIOINFORMATICS, 2016, 32 (12) : 37 - 43