# Quantile regression

Quantile regression is a type of regression analysis used in statistics and econometrics. Whereas the method of least squares results in estimates of the conditional mean of the response variable given certain values of the predictor variables, quantile regression aims at estimating either the conditional median or other quantiles of the response variable. Essentially, quantile regression is the extension of linear regression and we use it when the conditions of linear regression are not applicable.

Quantile regression is desired if conditional quantile functions are of interest. One advantage of quantile regression, relative to the ordinary least squares regression, is that the quantile regression estimates are more robust against outliers in the response measurements. However, the main attraction of quantile regression goes beyond that. Different measures of central tendency and statistical dispersion can be useful to obtain a more comprehensive analysis of the relationship between variables.[1]

In ecology, quantile regression has been proposed and used as a way to discover more useful predictive relationships between variables in cases where there is no relationship or only a weak relationship between the means of such variables. The need for and success of quantile regression in ecology has been attributed to the complexity of interactions between different factors leading to data with unequal variation of one variable for different ranges of another variable.[2]

Another application of quantile regression is in the areas of growth charts, where percentile curves are commonly used to screen for abnormal growth.[3][4]

## Mathematics

The mathematical forms arising from quantile regression are distinct from those arising in the method of least squares. The method of least squares leads to a consideration of problems in an inner product space, involving projection onto subspaces, and thus the problem of minimizing the squared errors can be reduced to a problem in numerical linear algebra. Quantile regression does not have this structure, and instead leads to problems in linear programming that can be solved by the simplex method.

## History

The idea of estimating a median regression slope, a major theorem about minimizing sum of the absolute deviances and a geometrical algorithm for constructing median regression was proposed in 1760 by Ruđer Josip Bošković, a Jesuit Catholic priest from Dubrovnik.[1]:4[5] He was interested in the ellipticity of the earth, building on Isaac Newton's suggestion that its rotation could cause it to bulge at the equator with a corresponding flattening at the poles.[6] He finally produced the first geometric procedure for determining the equator of a rotating planet from three observations of a surface feature. More importantly for quantile regression, he was able to develop the first evidence of the least absolute criterion and preceded the least squares introduced by Legendre in 1805 by fifty years.[7]

Other thinkers began building upon Bošković's idea such as Pierre-Simon Laplace, who developed the so-called "methode de situation." This led to Francis Edgeworth's plural median[8] - a geometric approach to median regression - and is recognized as the precursor of the simplex method.[7] The works of Bošković, Laplace, and Edgeworth were recognized as a prelude to Roger Koenker's contributions to quantile regression.

Median regression computations for larger data sets are quite tedious compared to the least squares method, for which reason it has historically generated a lack of popularity among statisticians, until the widespread adoption of computers in the latter part of the 20th century.

## Quantiles

Let ${\displaystyle Y}$ be a real valued random variable with cumulative distribution function ${\displaystyle F_{Y}(y)=P(Y\leq y)}$. The ${\displaystyle \tau }$th quantile of Y is given by

${\displaystyle Q_{Y}(\tau )=F_{Y}^{-1}(\tau )=\inf \left\{y:F_{Y}(y)\geq \tau \right\}}$

where ${\displaystyle \tau \in (0,1).}$

Define the loss function as ${\displaystyle \rho _{\tau }(y)=y(\tau -\mathbb {I} _{(y<0)})}$, where ${\displaystyle \mathbb {I} }$ is an indicator function. A specific quantile can be found by minimizing the expected loss of ${\displaystyle Y-u}$ with respect to ${\displaystyle u}$:[1]:5–6

${\displaystyle {\underset {u}{\min }}E(\rho _{\tau }(Y-u))={\underset {u}{\min }}\left\{(\tau -1)\int _{-\infty }^{u}(y-u)dF_{Y}(y)+\tau \int _{u}^{\infty }(y-u)dF_{Y}(y)\right\}.}$

This can be shown by setting the derivative of the expected loss function to 0 and letting ${\displaystyle q_{\tau }}$ be the solution of

${\displaystyle 0=(1-\tau )\int _{-\infty }^{q_{\tau }}dF_{Y}(y)-\tau \int _{q_{\tau }}^{\infty }dF_{Y}(y).}$

This equation reduces to

${\displaystyle 0=F_{Y}(q_{\tau })-\tau ,}$

and then to

${\displaystyle F_{Y}(q_{\tau })=\tau .}$

Hence ${\displaystyle q_{\tau }}$ is ${\displaystyle \tau }$th quantile of the random variable Y.

## Example

Let ${\displaystyle Y}$ be a discrete random variable that takes values 1,2,..,9 with equal probabilities. The task is to find the median of Y, and hence the value ${\displaystyle \tau =0.5}$ is chosen. The expected loss, L(u), is

${\displaystyle L(u)={\frac {(\tau -1)}{9}}\sum _{y_{i}

Since ${\displaystyle {0.5/9}}$ is a constant, it can be taken out of the expected loss function (this is only true if ${\displaystyle \tau =0.5}$). Then, at u=3,

${\displaystyle L(3)\propto \sum _{i=1}^{2}-(i-3)+\sum _{i=3}^{9}(i-3)=[(2+1)+(0+1+2+...+6)]=24.}$

Suppose that u is increased by 1 unit. Then the expected loss will be changed by ${\displaystyle (3)-(6)=-3}$ on changing u to 4. If, u=5, the expected loss is

${\displaystyle L(5)\propto \sum _{i=1}^{4}i+\sum _{i=0}^{4}i=20,}$

and any change in u will increase the expected loss. Thus u=5 is the median. The Table below shows the expected loss (divided by ${\displaystyle {0.5/9}}$) for different values of u.

 u 1 2 3 4 5 6 7 8 9 Expected loss 36 29 24 21 20 21 24 29 36

## Intuition

Consider ${\displaystyle \tau =0.5}$ and let q be an initial guess for ${\displaystyle q_{\tau }}$. The expected loss evaluated at q is

${\displaystyle L(q)=-0.5\int _{-\infty }^{q}(y-q)dF_{Y}(y)+0.5\int _{q}^{\infty }(y-q)dF_{Y}(y).}$

In order to minimize the expected loss, we move the value of q a little bit to see whether the expected loss will rise or fall. Suppose we increase q by 1 unit. Then the change of expected loss would be

${\displaystyle \int _{-\infty }^{q}1dF_{Y}(y)-\int _{q}^{\infty }1dF_{Y}(y).}$

The first term of the equation is ${\displaystyle F_{Y}(q)}$ and second term of the equation is ${\displaystyle 1-F_{Y}(q)}$. Therefore, the change of expected loss function is negative if and only if ${\displaystyle F_{Y}(q)<0.5}$, that is if and only if q is smaller than the median. Similarly, if we reduce q by 1 unit, the change of expected loss function is negative if and only if q is larger than the median.

In order to minimize the expected loss function, we would increase (decrease) L(q) if q is smaller (larger) than the median, until q reaches the median. The idea behind the minimization is to count the number of points (weighted with the density) that are larger or smaller than q and then move q to a point where q is larger than ${\displaystyle 100\tau }$% of the points.

## Sample quantile

The ${\displaystyle \tau }$ sample quantile can be obtained by solving the following minimization problem

${\displaystyle {\hat {q}}_{\tau }={\underset {q\in \mathbb {R} }{\mbox{arg min}}}\sum _{i=1}^{n}\rho _{\tau }(y_{i}-q),}$
${\displaystyle ={\underset {q\in \mathbb {R} }{\mbox{arg min}}}\left[(\tau -1)\sum _{y_{i}, where the function ${\displaystyle \rho _{\tau }}$ is the tilted absolute value function. The intuition is the same as for the population quantile.

## Conditional quantile and quantile regression

Suppose the ${\displaystyle \tau }$th conditional quantile function is ${\displaystyle Q_{Y|X}(\tau )=X\beta _{\tau }}$. Given the distribution function of ${\displaystyle Y}$, ${\displaystyle \beta _{\tau }}$ can be obtained by solving

${\displaystyle \beta _{\tau }={\underset {\beta \in \mathbb {R} ^{k}}{\mbox{arg min}}}E(\rho _{\tau }(Y-X\beta )).}$

Solving the sample analog gives the estimator of ${\displaystyle \beta }$.

${\displaystyle {\hat {\beta _{\tau }}}={\underset {\beta \in \mathbb {R} ^{k}}{\mbox{arg min}}}\sum _{i=1}^{n}(\rho _{\tau }(Y_{i}-X_{i}\beta )).}$

## Computation

The minimization problem can be reformulated as a linear programming problem

${\displaystyle {\underset {\beta ,u^{+},u^{-}\in \mathbb {R} ^{k}\times \mathbb {R} _{+}^{2n}}{\min }}\left\{\tau 1_{n}^{'}u^{+}+(1-\tau )1_{n}^{'}u^{-}|X\beta +u^{+}-u^{-}=Y\right\},}$

where

${\displaystyle u_{j}^{+}=\max(u_{j},0)}$ ,    ${\displaystyle u_{j}^{-}=-\min(u_{j},0).}$

Simplex methods[1]:181 or interior point methods[1]:190 can be applied to solve the linear programming problem.

## Asymptotic properties

For ${\displaystyle \tau \in (0,1)}$, under some regularity conditions, ${\displaystyle {\hat {\beta }}_{\tau }}$ is asymptotically normal:

${\displaystyle {\sqrt {n}}({\hat {\beta }}_{\tau }-\beta _{\tau }){\overset {d}{\rightarrow }}N(0,\tau (1-\tau )D^{-1}\Omega _{x}D^{-1}),}$

where

${\displaystyle D=E(f_{Y}(X\beta )XX^{\prime })}$ and ${\displaystyle \Omega _{x}=E(X^{\prime }X).}$

Direct estimation of the asymptotic variance-covariance matrix is not always satisfactory. Inference for quantile regression parameters can be made with the regression rank-score tests or with the bootstrap methods.[9]

## Equivariance

See invariant estimator for background on invariance or see equivariance.

### Scale equivariance

For any ${\displaystyle a>0}$ and ${\displaystyle \tau \in [0,1]}$

${\displaystyle {\hat {\beta }}(\tau ;aY,X)=a{\hat {\beta }}(\tau ;Y,X),}$
${\displaystyle {\hat {\beta }}(\tau ;-aY,X)=-a{\hat {\beta }}(1-\tau ;Y,X).}$

### Shift equivariance

For any ${\displaystyle \gamma \in R^{k}}$ and ${\displaystyle \tau \in [0,1]}$

${\displaystyle {\hat {\beta }}(\tau ;Y+X\gamma ,X)={\hat {\beta }}(\tau ;Y,X)+\gamma .}$

### Equivariance to reparameterization of design

Let ${\displaystyle A}$ be any ${\displaystyle p\times p}$ nonsingular matrix and ${\displaystyle \tau \in [0,1]}$

${\displaystyle {\hat {\beta }}(\tau ;Y,XA)=A^{-1}{\hat {\beta }}(\tau ;Y,X).}$

### Invariance to monotone transformations

If ${\displaystyle h}$ is a nondecreasing function on 'R, the following invariance property applies:

${\displaystyle h(Q_{Y|X}(\tau ))\equiv Q_{h(Y)|X}(\tau ).}$

Example (1):

If ${\displaystyle W=\exp(Y)}$ and ${\displaystyle Q_{Y|X}(\tau )=X\beta _{\tau }}$, then ${\displaystyle Q_{W|X}(\tau )=\exp(X\beta _{\tau })}$. The mean regression does not have the same property since ${\displaystyle \operatorname {E} (\ln(Y))\neq \ln(\operatorname {E} (Y)).}$

## Bayesian methods for quantile regression

Because quantile regression does not normally assume a parametric likelihood for the conditional distributions of Y|X, the Bayesian methods work with a working likelihood. A convenient choice is the asymmetric Laplacian likelihood,[10] because the mode of the resulting posterior under a flat prior is the usual quantile regression estimates. The posterior inference, however, must be interpreted with care. Yang, Wang and He[11] provided a posterior variance adjustment for valid inference. In addition, Yang and He[12] showed that one can have asymptotically valid posterior inference if the working likelihood is chosen to be the empirical likelihood.

## Censored quantile regression

If the response variable is subject to censoring, the conditional mean is not identifiable without additional distributional assumptions, but the conditional quantile is often identifiable. For recent work on censored quantile regression, see: Portnoy[13] and Wang and Wang[14]

Example (2):

Let ${\displaystyle Y^{c}=\max(0,Y)}$ and ${\displaystyle Q_{Y|X}=X\beta _{\tau }}$. Then ${\displaystyle Q_{Y^{c}|X}(\tau )=\max(0,X\beta _{\tau })}$. This is the censored quantile regression model: estimated values can be obtained without making any distributional assumptions, but at the cost of computational difficulty,[15] some of which can be avoided by using a simple three step censored quantile regression procedure as an approximation.[16]

For random censoring on the response variables, the censored quantile regression of Portnoy (2003)[13] provides consistent estimates of all identifiable quantile functions based on reweighting each censored point appropriately.

## Implementations

Numerous statistical software packages include implementations of quantile regression:

• Matlab function quantreg[17]
• Eviews, since version 6.[citation needed]
• gretl has the quantreg command.[18]
• R offers several packages that implement quantile regression, most notably quantreg by Roger Koenker,[19] but also gbm,[20] quantregForest[21], qrnn[22] and qgam[23]
• Python, via Scikit-garden[24]
• SAS through proc quantreg (ver. 9.2) and proc quantselect (ver. 9.3).[25]
• Stata, via the qreg command.[26][27]
• Vowpal Wabbit, via --loss_function quantile.[28]
• Statsmodels package for Python, via QuantReg[29]
• Mathematica package QuantileRegression.m[30] hosted at the MathematicaForPrediction project at GitHub.

## References

1. Koenker, Roger (2005). Quantile Regression. Cambridge University Press. pp. 146–7. ISBN 978-0-521-60827-5.
2. ^ Cade, Brian S.; Noon, Barry R. (2003). "A gentle introduction to quantile regression for ecologists" (PDF). Frontiers in Ecology and the Environment. 1 (8): 412–420. doi:10.2307/3868138. JSTOR 3868138.
3. ^ Wei, Y.; Pere, A.; Koenker, R.; He, X. (2006). "Quantile Regression Methods for Reference Growth Charts". Statistics in Medicine. 25 (8): 1369–1382. doi:10.1002/sim.2271. PMID 16143984.
4. ^ Wei, Y.; He, X. (2006). "Conditional Growth Charts (with discussions)". Annals of Statistics. 34 (5): 2069–2097 and 2126–2131. arXiv:math/0702634. doi:10.1214/009053606000000623.
5. ^ Stigler, S. (1984). "Boscovich, Simpson and a 1760 manuscript note on fitting a linear relation". Biometrika. 71 (3): 615–620. doi:10.1093/biomet/71.3.615.
6. ^ Koenker, Roger (2005). Quantile Regression. Cambridge: Cambridge University Press. p. 2. ISBN 9780521845731.
7. ^ a b Furno, Marilena; Vistocco, Domenico (2018). Quantile Regression: Estimation and Simulation. Hoboken, NJ: John Wiley & Sons. pp. xv. ISBN 9781119975281.
8. ^ Koenker, Roger (August 1998). "Galton, Edgeworth, Frisch, and prospects for quantile regression in economics" (PDF). UIUC.edu. Retrieved August 22, 2018.
9. ^ Kocherginsky, M.; He, X.; Mu, Y. (2005). "Practical Confidence Intervals for Regression Quantiles". Journal of Computational and Graphical Statistics. 14 (1): 41–55. doi:10.1198/106186005X27563.
10. ^ Kozumi, H.; Kobayashi, G. (2011). "Gibbs sampling methods for Bayesian quantile regression" (PDF). Journal of Statistical Computation and Simulation. 81 (11): 1565–1578. doi:10.1080/00949655.2010.496117.
11. ^ Yang, Y.; Wang, H.X.; He, X. (2016). "Posterior Inference in Bayesian Quantile Regression with Asymmetric Laplace Likelihood". International Statistical Review. 84 (3): 327–344. doi:10.1111/insr.12114.
12. ^ Yang, Y.; He, X. (2010). "Bayesian empirical likelihood for quantile regression". Annals of Statistics. 40 (2): 1102–1131. arXiv:1207.5378. doi:10.1214/12-AOS1005.
13. ^ a b Portnoy, S. L. (2003). "Censored Regression Quantiles". Journal of the American Statistical Association. 98 (464): 1001–1012. doi:10.1198/016214503000000954.
14. ^ Wang, H.; Wang, L. (2009). "Locally Weighted Censored Quantile Regression". Journal of the American Statistical Association. 104 (487): 1117–1128. CiteSeerX 10.1.1.504.796. doi:10.1198/jasa.2009.tm08230.
15. ^ Powell, James L. (1986). "Censored Regression Quantiles". Journal of Econometrics. 32 (1): 143–155. doi:10.1016/0304-4076(86)90016-3.
16. ^ Chernozhukov, Victor; Hong, Han (2002). "Three-Step Censored Quantile Regression and Extramarital Affairs". J. Amer. Statist. Assoc. 97 (459): 872–882. doi:10.1198/016214502388618663.
17. ^ "quantreg(x,y,tau,order,Nboot) - File Exchange - MATLAB Central". www.mathworks.com. Retrieved 2016-02-01.
18. ^ "Gretl Command Reference" (PDF). April 2017.
19. ^ "quantreg: Quantile Regression". R Project. 2018-12-18.
20. ^ "gbm: Generalized Boosted Regression Models". R Project. 2019-01-14.
21. ^ "quantregForest: Quantile Regression Forests". R Project. 2017-12-19.
22. ^ "qrnn: Quantile Regression Neural Networks". R Project. 2018-06-26.
23. ^ "qgam: Smooth Additive Quantile Regression Models". R Project. 2019-05-23.
24. ^ "Quantile Regression Forests". Scikit-garden. Retrieved 3 January 2019.
25. ^ "An Introduction to Quantile Regression and the QUANTREG Procedure" (PDF). SAS Support.
26. ^ "qreg — Quantile regression" (PDF). Stata Manual.
27. ^ Cameron, A. Colin; Trivedi, Pravin K. (2010). "Quantile Regression". Microeconometrics Using Stata (Revised ed.). College Station: Stata Press. pp. 211–234. ISBN 978-1-59718-073-3.
28. ^ "JohnLangford/vowpal_wabbit". GitHub. Retrieved 2016-07-09.
29. ^ "Quantile regression". statsmodels.org. Retrieved 3 January 2019.
30. ^ "QuantileRegression.m". MathematicaForPrediction. Retrieved 3 January 2019.