Smoothed Quantile Regression with Large-Scale Inference

12/09/2020
by   Xuming He, et al.
0

Quantile regression is a powerful tool for learning the relationship between a scalar response and a multivariate predictor in the presence of heavier tails and/or data heterogeneity. In the present paper, we consider statistical inference for quantile regression with large-scale data in the increasing dimension regime. We provide a comprehensive study of a convolution-type smoothing approach to achieving an adequate approximation to computation and inference for quantile regression. The ensuing estimator, which we refer to as conquer, turns the non-differentiable quantile loss function into a twice-differentiable, globally convex, and locally strongly convex surrogate, which admits a fast and scalable Barzilai-Borwein gradient-based algorithm to perform optimization, and a Rademacher multiplier bootstrap method for statistical inference. In the theoretical investigations of the conquer estimator, we establish nonasymptotic error bounds on the Bahadur-Kiefer linearization, from which we show that the asymptotic normality of the smoothed quantile regression estimator holds under a weaker requirement on the dimension of the predictors than needed for the exact quantile regression estimator. Our numerical studies confirm the conquer estimator as a practical and reliable approach to large-scale inference for quantile regression.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset
Success!
Error Icon An error occurred

Sign in with Google

×

Use your Google Account to sign in to DeepAI

×

Consider DeepAI Pro