Distributed inference for quantile regression processes
From MaRDI portal
Abstract: The increased availability of massive data sets provides a unique opportunity to discover subtle patterns in their distributions, but also imposes overwhelming computational challenges. To fully utilize the information contained in big data, we propose a two-step procedure: (i) estimate conditional quantile functions at different levels in a parallel computing environment; (ii) construct a conditional quantile regression process through projection based on these estimated quantile curves. Our general quantile regression framework covers both linear models with fixed or growing dimension and series approximation models. We prove that the proposed procedure does not sacrifice any statistical inferential accuracy provided that the number of distributed computing units and quantile levels are chosen properly. In particular, a sharp upper bound for the former and a sharp lower bound for the latter are derived to capture the minimal computational cost from a statistical perspective. As an important application, the statistical inference on conditional distribution functions is considered. Moreover, we propose computationally efficient approaches to conducting inference in the distributed estimation setting described above. Those approaches directly utilize the availability of estimators from sub-samples and can be carried out at almost no additional computational cost. Simulations confirm our statistical inferential theory.
Recommendations
Cited in
(67)- Smoothing Quantile Regressions
- Communication-efficient distributed estimation of partially linear additive models for large-scale data
- Communication-Efficient Distributed Linear Discriminant Analysis for Binary Classification
- Approximating Partial Likelihood Estimators via Optimal Subsampling
- Distributed subdata selection for big data via sampling-based approach
- Divide and recombine (D{\&}R) data science projects for deep analysis of big data and high computational complexity
- Nonparametric inference on smoothed quantile regression process
- A lack-of-fit test for quantile regression process models
- Functional linear quantile regression on a two-dimensional domain
- Distributed quantile regression for longitudinal big data
- Distributed Censored Quantile Regression
- Circumventing superefficiency: an effective strategy for distributed computing in non-standard problems
- A partitioned quasi-likelihood for distributed statistical inference
- Direct regression modelling of high-order moments in big data
- First-Order Newton-Type Estimator for Distributed Estimation and Inference
- High-dimensional integrative analysis with homogeneity and sparsity recovery
- Bayesian inference for quantile autoregressive model with explanatory variables
- Optimal subsampling for large‐sample quantile regression with massive data
- Communication-efficient estimation of high-dimensional quantile regression
- Communication-efficient distributed \(M\)-estimation with missing data
- Distributed optimal subsampling for quantile regression with massive data
- Communication-efficient estimation and inference for high-dimensional quantile regression based on smoothed decorrelated score
- Two-stage communication-efficient distributed sparse M-estimation with missing data
- Optimal subsample selection for massive logistic regression with distributed data
- Robust reduced rank regression in a distributed setting
- Distributed inference for two‐sample U‐statistics in massive data analysis
- Multi-dimensional latent group structures with heterogeneous distributions
- Robust communication-efficient distributed composite quantile regression and variable selection for massive data
- Discussion on the paper ‘A review of distributed statistical inference’
- Quantile regression in big data: a divide and conquer based strategy
- Distributed estimation and inference for semiparametric binary response models
- Distributed Estimation for Principal Component Analysis: An Enlarged Eigenspace Analysis
- A framework for mediation analysis with massive data
- Optimal subsampling algorithms for composite quantile regression in massive data
- An asymptotic analysis of distributed nonparametric methods
- Distributed linear regression by averaging
- Learning sparse conditional distribution: an efficient kernel-based approach
- Automatic variable selection in a linear model on massive data
- Optimal Subsampling Bootstrap for Massive Data
- Renewable composite quantile method and algorithm for nonparametric models with streaming data
- Global debiased DC estimations for biased estimators via pro forma regression
- A selective review on statistical methods for massive data computation: distributed computing, subsampling, and minibatch techniques
- Adapting the Hill estimator to distributed inference: dealing with the bias
- Distributed Sparse Composite Quantile Regression in Ultrahigh Dimensions
- Quantile regression under memory constraint
- Flexible specification testing in quantile regression models
- Multi-round smoothed composite quantile regression for distributed data
- Distributed statistical inference for massive data
- Residual projection for quantile regression in vertically partitioned big data
- Block average quantile regression for massive dataset
- Adaptive distributed inference for multi-source massive heterogeneous data
- Distributed adaptive Huber regression
- Distributed learning for sketched kernel regression
- Efficient fused learning for distributed imbalanced data
- On the unbiased asymptotic normality of quantile regression with fixed effects
- Communication-efficient sparse composite quantile regression for distributed data
- Quantile regression for large-scale applications
- Penalized Quantile Regression for Distributed Big Data Using the Slack Variable Representation
- Communication-Efficient Accurate Statistical Estimation
- A review of distributed statistical inference
- Parallel inference for massive distributed spatial data using low-rank models
- Divide and conquer for accelerated failure time model with massive time‐to‐event data
- Sampling-based estimation for massive survival data with additive hazards model
- A distributed quantile estimation algorithm of heavy-tailed distribution with massive datasets
- Distributed smoothed rank regression with heterogeneous errors for massive data
- Adaptive distributed support vector regression of massive data
- WONDER: weighted one-shot distributed ridge regression in high dimensions
This page was built for publication: Distributed inference for quantile regression processes
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2414100)