Orthogonalized Kernel Debiased Machine Learning for Multimodal Data Analysis
From MaRDI portal
Publication:6077573
DOI10.1080/01621459.2021.2013851arXiv2103.07088OpenAlexW3215001290MaRDI QIDQ6077573FDOQ6077573
Authors: Xiaowu Dai, Lexin Li
Publication date: 18 October 2023
Published in: Journal of the American Statistical Association (Search for Journal in Brave)
Abstract: Multimodal imaging has transformed neuroscience research. While it presents unprecedented opportunities, it also imposes serious challenges. Particularly, it is difficult to combine the merits of the interpretability attributed to a simple association model with the flexibility achieved by a highly adaptive nonlinear model. In this article, we propose an orthogonalized kernel debiased machine learning approach, which is built upon the Neyman orthogonality and a form of decomposition orthogonality, for multimodal data analysis. We target the setting that naturally arises in almost all multimodal studies, where there is a primary modality of interest, plus additional auxiliary modalities. We establish the root--consistency and asymptotic normality of the estimated primary parameter, the semi-parametric estimation efficiency, and the asymptotic validity of the confidence band of the predicted primary modality effect. Our proposal enjoys, to a good extent, both model interpretability and model flexibility. It is also considerably different from the existing statistical methods for multimodal data integration, as well as the orthogonality-based methods for high-dimensional inferences. We demonstrate the efficacy of our method through both simulations and an application to a multimodal neuroimaging study of Alzheimer's disease.
Full work available at URL: https://arxiv.org/abs/2103.07088
high-dimensional inferencereproducing kernel Hilbert spacebasis expansionNeyman orthogonalityneuroimaging analysismultimodal data integration
Cites Work
- Title not available (Why is that?)
- Asymptotic Statistics
- Statistics for high-dimensional data. Methods, theory and applications.
- Confidence intervals for high-dimensional linear regression: minimax rates and adaptivity
- Simultaneous analysis of Lasso and Dantzig selector
- Title not available (Why is that?)
- Sure Independence Screening for Ultrahigh Dimensional Feature Space
- Double/debiased machine learning for treatment and structural parameters
- Confidence Intervals for Low Dimensional Parameters in High Dimensional Linear Models
- On asymptotically optimal confidence regions and tests for high-dimensional models
- Statistical modeling: The two cultures. (With comments and a rejoinder).
- Joint and individual variation explained (JIVE) for integrated analysis of multiple data types
- Semiparametric efficiency bounds
- Title not available (Why is that?)
- Bayesian generalized low rank regression models for neuroimaging phenotypes and genetic markers
- Causality. Models, reasoning, and inference
- Minimax Rates of Estimation for High-Dimensional Linear Regression Over $\ell_q$-Balls
- Estimation and model selection in generalized additive partial linear models for correlated data with diverging number of covariates
- Improved rates and asymptotic normality for nonparametric neural network estimators
- Component selection and smoothing in multivariate nonparametric regression
- Efficient Estimation in Marginal Partially Linear Models for Longitudinal/Clustered Data Using Splines
- Anti-concentration and honest, adaptive confidence bands
- Semiparametric Efficiency in Multivariate Regression Models with Missing Data
- An exponential inequality for the distribution function of the kernel density estimator, with applications to adaptive estimation
- Linear smoothers and additive models
- Analysis of a random forests model
- Title not available (Why is that?)
- Nonparametric Inferences for Additive Models
- Title not available (Why is that?)
- Title not available (Why is that?)
- A review of statistical methods in imaging genetics
- Estimation and inference in generalized additive coefficient models for nonlinear interactions with high-dimensional covariates
- D-CCA: A Decomposition-Based Canonical Correlation Analysis for High-Dimensional Datasets
- An iterative penalized least squares approach to sparse canonical correlation analysis
- Kernel meets sieve: post-regularization confidence bands for sparse additive model
- Integrating Multisource Block-Wise Missing Data in Model Selection
- Integrative multi-view regression: bridging group-sparse and low-rank models
- Integrative Factor Regression and Its Inference for Multimodal Data Analysis
Cited In (2)
This page was built for publication: Orthogonalized Kernel Debiased Machine Learning for Multimodal Data Analysis
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6077573)