Generalized matrix decomposition regression: estimation and inference for two-way structured data
From MaRDI portal
Publication:6138615
Abstract: This paper studies high-dimensional regression with two-way structured data. To estimate the high-dimensional coefficient vector, we propose the generalized matrix decomposition regression (GMDR) to efficiently leverage any auxiliary information on row and column structures. The GMDR extends the principal component regression (PCR) to two-way structured data, but unlike PCR, the GMDR selects the components that are most predictive of the outcome, leading to more accurate prediction. For inference on regression coefficients of individual variables, we propose the generalized matrix decomposition inference (GMDI), a general high-dimensional inferential framework for a large family of estimators that include the proposed GMDR estimator. GMDI provides more flexibility for modeling relevant auxiliary row and column structures. As a result, GMDI does not require the true regression coefficients to be sparse; it also allows dependent and heteroscedastic observations. We study the theoretical properties of GMDI in terms of both the type-I error rate and power and demonstrate the effectiveness of GMDR and GMDI on simulation studies and an application to human microbiome data.
Cites work
- scientific article; zbMATH DE number 845714 (Why is no real title available?)
- scientific article; zbMATH DE number 6159604 (Why is no real title available?)
- A fast small‐sample kernel independence test for microbiome community‐level association analysis
- A general theory of hypothesis tests and confidence regions for sparse high dimensional models
- A generalized least-square matrix decomposition
- A significance test for graph-constrained estimation
- Confidence Intervals and Hypothesis Testing for High-Dimensional Regression
- Confidence intervals for low dimensional parameters in high dimensional linear models
- Estimating the error variance in a high-dimensional linear model
- Fisher lecture: Dimension reduction in regression
- High-dimensional statistics. A non-asymptotic viewpoint
- Hypothesis Testing in High-Dimensional Regression Under the Gaussian Random Design Model: Asymptotic Theory
- Inference for High-Dimensional Linear Mixed-Effects Models: A Quasi-Likelihood Approach
- Kernel-penalized regression for analysis of microbiome data
- Linear hypothesis testing in dense high-dimensional linear models
- On asymptotically optimal confidence regions and tests for high-dimensional models
- On the conditions used to prove oracle results for the Lasso
- Operator related to a data matrix: a survey
- Scaled sparse linear regression
- Semiparametric Regression of Multidimensional Genetic Pathway Data: Least‐Squares Kernel Machines and Linear Mixed Models
- Statistical significance in high-dimensional linear models
- The benefit of group sparsity in group inference with de-biased scaled group Lasso
- The control of the false discovery rate in multiple testing under dependency.
- The elements of statistical learning. Data mining, inference, and prediction
- The sparsity and bias of the LASSO selection in high-dimensional linear regression
- Uniform post-selection inference for least absolute deviation regression and other Z-estimation problems
This page was built for publication: Generalized matrix decomposition regression: estimation and inference for two-way structured data
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6138615)