Kernel dimension reduction in regression

From MaRDI portal
Publication:2388983

DOI10.1214/08-AOS637zbMath1168.62049arXiv0908.1854OpenAlexW2072460512MaRDI QIDQ2388983

Kenji Fukumizu, Michael I. Jordan, Francis Bach

Publication date: 22 July 2009

Published in: The Annals of Statistics (Search for Journal in Brave)

Full work available at URL: https://arxiv.org/abs/0908.1854



Related Items

A selective overview of sparse sufficient dimension reduction, Entropy-based sliced inverse regression, A Rigorous Theory of Conditional Mean Embeddings, Testing the Linear Mean and Constant Variance Conditions in Sufficient Dimension Reduction, Sufficient dimension reduction via principal L\(q\) support vector machine, A brief review of linear sufficient dimension reduction through optimization, Sliced Inverse Regression in Metric Spaces, Rejoinder on: ``Probability enhanced effective dimension reduction for classifying sparse functional data, Functional sufficient dimension reduction through average Fréchet derivatives, High-dimensional sufficient dimension reduction through principal projections, Minimal \(\sigma\)-field for flexible sufficient dimension reduction, Nonlinear predictive directions in clinical trials, Discriminant Dynamic Mode Decomposition for Labeled Spatiotemporal Data Collections, Student sliced inverse regression, Canonical kernel dimension reduction, Probability-enhanced sufficient dimension reduction for binary classification, Unnamed Item, Nonlinear and additive principal component analysis for functional data, Central subspaces review: methods and applications, Sufficient Dimension Reduction via Squared-Loss Mutual Information Estimation, Sparse SIR: optimal rates and adaptive estimation, Predictive power of principal components for single-index model and sufficient dimension reduction, A new reproducing kernel‐based nonlinear dimension reduction method for survival data, A structured covariance ensemble for sufficient dimension reduction, An Outer-Product-of-Gradient Approach to Dimension Reduction and its Application to Classification in High Dimensional Space, Supervised distance preserving projection using alternating direction method of multipliers, Conditional Functional Graphical Models, Dimensionality reduction by mixed kernel canonical correlation analysis, Tail inverse regression: dimension reduction for prediction of extremes, An Additive Graphical Model for Discrete Data, Canonical dependency analysis based on squared-loss mutual information, Multivariate tests of independence based on a new class of measures of independence in reproducing kernel Hilbert space, Minimax risks for sparse regressions: ultra-high dimensional phenomenons, Learning sparse gradients for variable selection and dimension reduction, A general theory for nonlinear sufficient dimension reduction: formulation and estimation, Principal support vector machines for linear and nonlinear sufficient dimension reduction, Sufficient dimension reduction based on an ensemble of minimum average variance estimators, A study on imbalance support vector machine algorithms for sufficient dimension reduction, Dimension reduction and its application to model-based exploration in continuous spaces, Conditional Density Estimation with Dimensionality Reduction via Squared-Loss Conditional Entropy Minimization, Direct Estimation of the Derivative of Quadratic Mutual Information with Application in Supervised Dimension Reduction, An adaptive composite quantile approach to dimension reduction, Sufficient dimension reduction with simultaneous estimation of effective dimensions for time-to-event data, Information-Theoretic Semi-Supervised Metric Learning via Entropy Regularization, Minimax adaptive dimension reduction for regression, Dealing with big data: comparing dimension reduction and shrinkage regression methods, Least-squares independence regression for non-linear causal inference under non-Gaussian noise, Sufficient dimension reduction and prediction in regression, Dimensionality reduction: an interpretation from manifold regularization perspective, Sufficient Dimension Reduction via Direct Estimation of the Gradients of Logarithmic Conditional Densities, Statistical performance of optimal scoring in reproducing kernel Hilbert spaces, Slice inverse regression with score functions, Characteristic and Universal Tensor Product Kernels, A class of optimal estimators for the covariance operator in reproducing kernel Hilbert spaces, Sparse Estimation of Conditional Graphical Models With Application to Gene Networks, Kernel sliced inverse regression: regularization and consistency, A Cost Based Reweighted Scheme of Principal Support Vector Machine, A Conditional Entropy Minimization Criterion for Dimensionality Reduction and Multiple Kernel Learning, Subspace perspective on canonical correlation analysis: dimension reduction and minimax rates, Estimating sufficient reductions of the predictors in abundant high-dimensional regressions, Model-based reinforcement learning with dimension reduction, On the predictive potential of kernel principal components, Computing functions of random variables via reproducing kernel Hilbert space representations, Comments on: ``Probability enhanced effective dimension reduction for classifying sparse functional data, Least-Squares Independent Component Analysis, A Shrinkage Estimation of Central Subspace in Sufficient Dimension Reduction, Comment, Machine learning with squared-loss mutual information, Isometric sliced inverse regression for nonlinear manifold learning, Gradient-Based Kernel Dimension Reduction for Regression, On an Additive Semigraphoid Model for Statistical Networks With Application to Pathway Analysis, Discussion of: Brownian distance covariance, Admissible kernels for RKHS embedding of probability distributions, Sliced inverse regression method for multivariate compositional data modeling, A Nonparametric Graphical Model for Functional Data With Application to Brain Networks Based on fMRI, A Generalized Kernel Method for Global Sensitivity Analysis



Cites Work