Large-Scale Sparse Regression for Multiple Responses with Applications to UK Biobank

In high-dimensional regression problems, often a relatively small subset of the features are relevant for predicting the outcome, and methods that impose sparsity on the solution are popular. When multiple correlated outcomes are available (multitask), reduced rank regression is an effective way to borrow strength and capture latent structures that underlie the data. Our proposal is motivated by the UK Biobank population-based cohort study, where we are faced with large-scale, ultrahigh-dimensional features, and have access to a large number of outcomes (phenotypes): lifestyle measures, biomarkers, and disease outcomes. We are hence led to fit sparse reduced-rank regression models, using computational strategies that allow us to scale to problems of this size. We use an iterative algorithm that alternates between solving the sparse regression problem and solving the reduced rank decomposition. For the sparse regression component, we propose a scalable iterative algorithm based on adaptive screening that leverages the sparsity assumption and enables us to focus on solving much smaller sub-problems. The full solution is reconstructed and tested via an optimality condition to make sure it is a valid solution for the original problem. We further extend the method to cope with practical issues such as the inclusion of confounding variables and imputation of missing values among the phenotypes. Experiments on both synthetic data and the UK Biobank data demonstrate the effectiveness of the method and the algorithm. We present multiSnpnet package, available at http://github.com/junyangq/multiSnpnet that works on top of PLINK2 files, which we anticipate to be a valuable tool for generating polygenic risk scores from human genetic studies.

[1]  Lin Xiao,et al.  Dual Averaging Methods for Regularized Stochastic Learning and Online Optimization , 2009, J. Mach. Learn. Res..

[2]  Fei Wang,et al.  Leveraging mixed and incomplete outcomes via reduced-rank modeling , 2018, J. Multivar. Anal..

[3]  T. W. Anderson Estimating Linear Restrictions on Regression Coefficients for Multivariate Normal Distributions , 1951 .

[4]  M. Rivas,et al.  Medical relevance of protein-truncating variants across 337,205 individuals in the UK Biobank study , 2018, Nature Communications.

[5]  Trevor Hastie,et al.  A Fast and Flexible Algorithm for Solving the Lasso in Large-scale and Ultrahigh-dimensional Problems , 2019 .

[6]  John C. Gower,et al.  Understanding Biplots: Gower/Understanding Biplots , 2011 .

[7]  Timo Similä,et al.  Input selection and shrinkage in multiresponse linear regression , 2007, Comput. Stat. Data Anal..

[8]  Trevor Hastie,et al.  Genetics of 35 blood and urine biomarkers in the UK Biobank , 2020, Nature Genetics.

[9]  P. Bickel,et al.  SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR , 2008, 0801.1095.

[10]  Michael I. Jordan,et al.  Union support recovery in high-dimensional multivariate regression , 2008, 2008 46th Annual Allerton Conference on Communication, Control, and Computing.

[11]  D. Reich,et al.  Principal components analysis corrects for stratification in genome-wide association studies , 2006, Nature Genetics.

[12]  Jianhua Z. Huang,et al.  Sparse Reduced-Rank Regression for Simultaneous Dimension Reduction and Variable Selection , 2012 .

[13]  P. Visscher,et al.  10 Years of GWAS Discovery: Biology, Function, and Translation. , 2017, American journal of human genetics.

[14]  Marco Cesati,et al.  Understanding the Linux Kernel - from I / O ports to process management: covers Linux Kernel version 2.4 (2. ed.) , 2005 .

[15]  G. Reinsel,et al.  Multivariate Reduced-Rank Regression: Theory and Applications , 1998 .

[16]  Martin J. Wainwright,et al.  Sharp Thresholds for High-Dimensional and Noisy Sparsity Recovery Using $\ell _{1}$ -Constrained Quadratic Programming (Lasso) , 2009, IEEE Transactions on Information Theory.

[17]  Thomas Viklands Algorithms for the Weighted Orthogonal Procrustes Problem and other Least Squares Problems , 2006 .

[18]  P. Donnelly,et al.  The UK Biobank resource with deep phenotyping and genomic data , 2018, Nature.

[19]  Léon Bottou,et al.  Large-Scale Machine Learning with Stochastic Gradient Descent , 2010, COMPSTAT.

[20]  Trevor Hastie,et al.  A fast and scalable framework for large-scale and ultrahigh-dimensional sparse regression with application to the UK Biobank , 2019, bioRxiv.

[21]  Scott Shenker,et al.  Spark: Cluster Computing with Working Sets , 2010, HotCloud.

[22]  Ji Zhu,et al.  Regularized Multivariate Regression for Identifying Master Predictors with Application to Integrative Genomics Study of Breast Cancer. , 2008, The annals of applied statistics.

[23]  Zhuang Ma,et al.  Adaptive Estimation in Two-way Sparse Reduced-rank Regression , 2014, 1403.1922.

[24]  Robert Tibshirani,et al.  Spectral Regularization Algorithms for Learning Large Incomplete Matrices , 2010, J. Mach. Learn. Res..

[25]  Walter Zucchini,et al.  Model Selection , 2011, International Encyclopedia of Statistical Science.

[26]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[27]  Trevor Hastie,et al.  Regularization Paths for Generalized Linear Models via Coordinate Descent. , 2010, Journal of statistical software.

[28]  M. Yuan,et al.  Model selection and estimation in regression with grouped variables , 2006 .

[29]  Sara van de Geer,et al.  Statistics for High-Dimensional Data: Methods, Theory and Applications , 2011 .

[30]  H. Zou,et al.  Regularization and variable selection via the elastic net , 2005 .

[31]  S. Keleş,et al.  Sparse partial least squares regression for simultaneous dimension reduction and variable selection , 2010, Journal of the Royal Statistical Society. Series B, Statistical methodology.

[32]  Matti Pirinen,et al.  MetaPhat: Detecting and Decomposing Multivariate Associations From Univariate Genome-Wide Association Statistics , 2019, bioRxiv.

[33]  S. O’Brien,et al.  SmileFinder: a resampling-based approach to evaluate signatures of selection from genome-wide sets of matching allele frequency data in two or more diploid populations , 2015, GigaScience.

[34]  Stephen J. Wright,et al.  Simultaneous Variable Selection , 2005, Technometrics.

[35]  Francis R. Bach,et al.  Consistency of the group Lasso and multiple kernel learning , 2007, J. Mach. Learn. Res..

[36]  Ab Mooijaart,et al.  A general solution of the weighted orthonormal procrustes problem , 1990 .

[37]  Trevor Hastie,et al.  Polygenic risk modeling with latent trait-related genetic components , 2019, European Journal of Human Genetics.

[38]  Nickolay T. Trendafilov,et al.  On a differential equation approach to the weighted orthogonal Procrustes problem , 1998, Stat. Comput..

[39]  Sanjay Ghemawat,et al.  MapReduce: simplified data processing on large clusters , 2008, CACM.

[40]  Trevor Hastie,et al.  Fast Lasso method for Large-scale and Ultrahigh-dimensional Cox Model with applications to UK Biobank , 2020, bioRxiv.

[41]  K. Gabriel,et al.  The biplot graphic display of matrices with application to principal component analysis , 1971 .

[42]  N. Meinshausen,et al.  High-dimensional graphs and variable selection with the Lasso , 2006, math/0608017.

[43]  Gustavo de los Campos,et al.  Accurate Genomic Prediction Of Human Height , 2017, bioRxiv.

[44]  R. Tibshirani,et al.  Strong rules for discarding predictors in lasso‐type problems , 2010, Journal of the Royal Statistical Society. Series B, Statistical methodology.

[45]  Sara van de Geer,et al.  Statistics for High-Dimensional Data , 2011 .

[46]  H. Hotelling Relations Between Two Sets of Variates , 1936 .

[47]  Xiaokang Liu,et al.  Integrative multi‐view regression: Bridging group‐sparse and low‐rank models , 2019, Biometrics.

[48]  Carson C Chow,et al.  Second-generation PLINK: rising to the challenge of larger and richer datasets , 2014, GigaScience.

[49]  S. Phinn,et al.  Australian vegetated coastal ecosystems as global hotspots for climate change mitigation , 2019, Nature Communications.

[50]  Guillaume Obozinski,et al.  Fast Algorithms for Sparse Reduced-Rank Regression , 2019, AISTATS.

[51]  Peng Zhao,et al.  On Model Selection Consistency of Lasso , 2006, J. Mach. Learn. Res..

[52]  Martin J. Wainwright,et al.  Dual Averaging for Distributed Optimization: Convergence Analysis and Network Scaling , 2010, IEEE Transactions on Automatic Control.

[53]  Y. Ritov,et al.  Persistence in high-dimensional linear predictor selection and the virtue of overparametrization , 2004 .