Orthogonalized Kernel Debiased Machine Learning for Multimodal Data Analysis
From MaRDI portal
Publication:6077573
Abstract: Multimodal imaging has transformed neuroscience research. While it presents unprecedented opportunities, it also imposes serious challenges. Particularly, it is difficult to combine the merits of the interpretability attributed to a simple association model with the flexibility achieved by a highly adaptive nonlinear model. In this article, we propose an orthogonalized kernel debiased machine learning approach, which is built upon the Neyman orthogonality and a form of decomposition orthogonality, for multimodal data analysis. We target the setting that naturally arises in almost all multimodal studies, where there is a primary modality of interest, plus additional auxiliary modalities. We establish the root--consistency and asymptotic normality of the estimated primary parameter, the semi-parametric estimation efficiency, and the asymptotic validity of the confidence band of the predicted primary modality effect. Our proposal enjoys, to a good extent, both model interpretability and model flexibility. It is also considerably different from the existing statistical methods for multimodal data integration, as well as the orthogonality-based methods for high-dimensional inferences. We demonstrate the efficacy of our method through both simulations and an application to a multimodal neuroimaging study of Alzheimer's disease.
Cites work
- scientific article; zbMATH DE number 3155150 (Why is no real title available?)
- scientific article; zbMATH DE number 3856278 (Why is no real title available?)
- scientific article; zbMATH DE number 3738700 (Why is no real title available?)
- scientific article; zbMATH DE number 45848 (Why is no real title available?)
- scientific article; zbMATH DE number 47282 (Why is no real title available?)
- scientific article; zbMATH DE number 720689 (Why is no real title available?)
- A review of statistical methods in imaging genetics
- An exponential inequality for the distribution function of the kernel density estimator, with applications to adaptive estimation
- An iterative penalized least squares approach to sparse canonical correlation analysis
- Analysis of a random forests model
- Anti-concentration and honest, adaptive confidence bands
- Asymptotic Statistics
- Bayesian generalized low rank regression models for neuroimaging phenotypes and genetic markers
- Causality. Models, reasoning, and inference
- Component selection and smoothing in multivariate nonparametric regression
- Confidence intervals for high-dimensional linear regression: minimax rates and adaptivity
- Confidence intervals for low dimensional parameters in high dimensional linear models
- D-CCA: A Decomposition-Based Canonical Correlation Analysis for High-Dimensional Datasets
- Double/debiased machine learning for treatment and structural parameters
- Efficient Estimation in Marginal Partially Linear Models for Longitudinal/Clustered Data Using Splines
- Estimation and inference in generalized additive coefficient models for nonlinear interactions with high-dimensional covariates
- Estimation and model selection in generalized additive partial linear models for correlated data with diverging number of covariates
- Improved rates and asymptotic normality for nonparametric neural network estimators
- Integrating Multisource Block-Wise Missing Data in Model Selection
- Integrative Factor Regression and Its Inference for Multimodal Data Analysis
- Integrative multi-view regression: bridging group-sparse and low-rank models
- Joint and individual variation explained (JIVE) for integrated analysis of multiple data types
- Kernel meets sieve: post-regularization confidence bands for sparse additive model
- Linear smoothers and additive models
- Minimax Rates of Estimation for High-Dimensional Linear Regression Over $\ell_q$-Balls
- Nonparametric Inferences for Additive Models
- On asymptotically optimal confidence regions and tests for high-dimensional models
- Semiparametric Efficiency in Multivariate Regression Models with Missing Data
- Semiparametric efficiency bounds
- Simultaneous analysis of Lasso and Dantzig selector
- Statistical modeling: The two cultures. (With comments and a rejoinder).
- Statistics for high-dimensional data. Methods, theory and applications.
- Sure independence screening for ultrahigh dimensional feature space. With discussion and authors' reply
Cited in
(2)
This page was built for publication: Orthogonalized Kernel Debiased Machine Learning for Multimodal Data Analysis
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6077573)