Bayesian variable selection with shrinking and diffusing priors
From MaRDI portal
Publication:118687
DOI10.1214/14-aos1207zbMath1302.62158arXiv1405.6545MaRDI QIDQ118687
Naveen Naidu Narisetty, Xuming He, Xuming He, Naveen Naidu Narisetty
Publication date: 1 April 2014
Published in: The Annals of Statistics (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1405.6545
Asymptotic properties of parametric estimators (62F12) Linear regression; mixed models (62J05) Bayesian inference (62F15)
Related Items
On the non-local priors for sparsity selection in high-dimensional Gaussian DAG models, Combining a relaxed EM algorithm with Occam's razor for Bayesian variable selection in high-dimensional regression, Strong selection consistency of Bayesian vector autoregressive models based on a pseudo-likelihood approach, Bayesian empirical likelihood inference and order shrinkage for autoregressive models, Comparison of Bayesian predictive methods for model selection, Bayesian Edge Regression in Undirected Graphical Models to Characterize Interpatient Heterogeneity in Cancer, Bayesian Regression Using a Prior on the Model Fit: The R2-D2 Shrinkage Prior, Bayesian inference in high-dimensional linear models using an empirical correlation-adaptive prior, Bayesian Estimation of Gaussian Conditional Random Fields, Bayesian high-dimensional semi-parametric inference beyond sub-Gaussian errors, Bayesian penalized Buckley-James method for high dimensional bivariate censored regression models, Objective Bayesian edge screening and structure selection for Ising networks, A generalized likelihood-based Bayesian approach for scalable joint regression and covariance selection in high dimensions, Random weighting in LASSO regression, Bayesian factor-adjusted sparse regression, Variable selection for high-dimensional genomic data with censored outcomes using group Lasso prior, High-dimensional multivariate posterior consistency under global-local shrinkage priors, Bayesian Approaches to Shrinkage and Sparse Estimation, Bayesian group selection in logistic regression with application to MRI data analysis, Identifying regions of inhomogeneities in spatial processes via an M‐RA and mixture priors, Projective inference in high-dimensional problems: prediction and feature selection, The EAS approach for graphical selection consistency in vector autoregression models, Weighted Bayesian bootstrap for scalable posterior distributions, A Bayesian Subset Specific Approach to Joint Selection of Multiple Graphical Models, On Joint Estimation of Gaussian Graphical Models for Spatial and Temporal Data, BAYESIAN DYNAMIC VARIABLE SELECTION IN HIGH DIMENSIONS, Bayesian analysis in single-index quantile regression with missing observation, Variable Selection Via Thompson Sampling, Neuronized Priors for Bayesian Sparse Linear Regression, A latent slice sampling algorithm, Nearly optimal Bayesian shrinkage for high-dimensional regression, An ensemble EM algorithm for Bayesian variable selection, A robust Bayesian analysis of variable selection under prior ignorance, The expectation-maximization approach for Bayesian additive Cox regression with current status data, The EAS approach to variable selection for multivariate response data in high-dimensional settings, Additive Bayesian variable selection under censoring and misspecification, Complexity analysis of Bayesian learning of high-dimensional DAG models and their equivalence classes, Bayesian Change Point Detection with Spike-and-Slab Priors, Consistent Bayesian sparsity selection for high-dimensional Gaussian DAG models with multiplicative and beta-mixture priors, Bayesian adaptive Lasso estimation of large graphical model based on modified Cholesky decomposition, A Mass-Shifting Phenomenon of Truncated Multivariate Normal Priors, Bayesian inference for an unknown number of attributes in restricted latent class models, The Bayesian nested Lasso for mixed frequency regression models, Unnamed Item, Bayesian multiple quantile regression for linear models using a score likelihood, Concentration of posterior model probabilities and normalized \({L_0}\) criteria, Bayesian empirical likelihood of quantile regression with missing observations, Bayesian fusion estimation via \(t\) shrinkage, Discussion, Fast Bayesian variable selection for high dimensional linear models: marginal solo spike and slab priors, Model selection using mass-nonlocal prior, Consistent group selection with Bayesian high dimensional modeling, Posterior graph selection and estimation consistency for high-dimensional Bayesian DAG models, Asymptotic theory of dependent Bayesian multiple testing procedures under possible model misspecification, Targeted Random Projection for Prediction From High-Dimensional Features, Maximum pairwise Bayes factors for covariance structure testing, Approximate Spectral Gaps for Markov Chain Mixing Times in High Dimensions, Joint Bayesian Variable and DAG Selection Consistency for High-dimensional Regression Models with Network-structured Covariates, Bayesian estimation of sparse signals with a continuous spike-and-slab prior, Variable selection for partially linear models via Bayesian subset modeling with diffusing prior, Quasi-Bayesian estimation of large Gaussian graphical models, Multivariate Bayesian variable selection exploiting dependence structure among outcomes: Application to air pollution effects on DNA methylation, Bayesian variable selection with shrinking and diffusing priors, basad, Minimax posterior convergence rates and model selection consistency in high-dimensional DAG models based on sparse Cholesky factors, Unified Bayesian theory of sparse linear regression with nuisance parameters, High-dimensional posterior consistency for hierarchical non-local priors in regression, Pruning variable selection ensembles, Variable selection consistency of Gaussian process regression, Bayesian estimation of large precision matrix based on Cholesky decomposition, Probabilistic Community Detection With Unknown Number of Communities, Skinny Gibbs: A Consistent and Scalable Gibbs Sampler for Model Selection, Bayesian Regularization for Graphical Models With Unequal Shrinkage, Adaptive random neighbourhood informed Markov chain Monte Carlo for high-dimensional Bayesian variable selection, Subjective Bayesian testing using calibrated prior probabilities, Particle EM for Variable Selection, Bayesian inference for high‐dimensional linear regression under mnet priors
Uses Software
Cites Work
- Bayesian variable selection with shrinking and diffusing priors
- The Adaptive Lasso and Its Oracle Properties
- Bayesian empirical likelihood for quantile regression
- Consistency of spike and slab regression
- A tail inequality for quadratic forms of subgaussian random vectors
- Consistency of objective Bayes factors as the model dimension grows
- Estimating the dimension of a model
- Optimal predictive model selection.
- Nonconcave penalized likelihood with a diverging number of parameters.
- Penalized orthogonal-components regression for large \(p\) small \(n\) data
- Spike and slab variable selection: frequentist and Bayesian strategies
- Bayesian variable selection for high dimensional generalized linear models: convergence rates of the fitted densities
- The Dantzig selector: statistical estimation when \(p\) is much larger than \(n\). (With discussions and rejoinder).
- Calibration and empirical Bayes variable selection
- A Selective Overview of Variable Selection in High Dimensional Feature Space (Invited Review Article)
- DASSO: Connections Between the Dantzig Selector and Lasso
- AN IN-DEPTH LOOK AT HIGHEST POSTERIOR MODEL SELECTION
- Bayesian Variable Selection in Linear Regression
- Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties
- Consistent High-Dimensional Bayesian Variable Selection via Penalized Credible Regions
- Likelihood-Based Selection and Sharp Parameter Estimation
- Bayesian Model Selection in High-Dimensional Settings
- Asymptotic oracle properties of SCAD-penalized least squares estimators
- Bayesian Subset Modeling for High-Dimensional Generalized Linear Models
- Simultaneous Regression Shrinkage, Variable Selection, and Supervised Clustering of Predictors with OSCAR
- Efficient Empirical Bayes Variable Selection and Estimation in Linear Models
- Unnamed Item
- Unnamed Item
- Unnamed Item