Playing with quantiles, part 1

March 8, 2011

(This article was first published on Freakonometrics - Tag - R-english, and kindly contributed to R-bloggers)

A standard idea in extreme value theory (see e.g. here,
in French unfortunately) is that to estimate the 99.5%
quantile (say), we just need to estimate a quantile of level 95% for
exceeding the 90% quantile.

In extreme value theory, we assume
that the 90% quantile (of the initial distribution) can be obtained easily, e.g. the empirical
quantile, and then, for the exceeding observations, we fit a Pareto
distribution (a Generalized Pareto one to be precise), and get a parametric quantile for the 95% quantile. I.e.

which can be written

So, an estimation of the cumulative distribution function is

and if we invert it, we get the popular expression for high level

Hence, we do not really care about observations in the core of the

And I was wondering if this can be transposed with quantile
regressions. Hence, I would like to get a quantile regression of level
90% (say) of
based on observations‘s,
but all
observations such that for
some are
missing. More precisely,
I have the following sample (here half of the observations are missing),

Assume that we know that I have observations below the
quantile of level 25%, and above the
quantile of level 75%.
If I want to get the 90% quantile regression, and the 10% quantile, the
code is simply,

Sigma <- matrix(c(1,r,r,1), 2, 2)
points(X[I],Y[I],col="light blue",cex=.7)

The graph is the following

Dotted lines – in black – are theoretical lines (if I had all
observations), and plain lines are (where half of the sample if
missing). Instead of a standard linear quantile regression, it is also
possible to try a spline regression,

So obviously, if I miss something in the middle, that’s no big deal, doted and plain lines are here extremely close.
But what if observations and
were correlated ? Consider a Gaussian random vector with
(here 0.6).

It looks like we overestimate the slope for high quantile, but not for
lower quantiles. So if observations are correlated, we have to be
cautious with that technique.
But why could that be interesting ? Well, because I wanted to run a
quantile regression on marathon results. But I could not get the
overall dataset (since I had to import observations manually, and I have to
admit that it was a bit boring). So I extracted finish times of the first
10% athletes, and the latest 10%. And I was
wondering if it was enough to look at the 5% and 95% quantiles, based
on the age of the runner… To be

To leave a comment for the author, please follow the link and comment on their blog: Freakonometrics - Tag - R-english. offers daily e-mail updates about R news and tutorials on topics such as: Data science, Big Data, R jobs, visualization (ggplot2, Boxplots, maps, animation), programming (RStudio, Sweave, LaTeX, SQL, Eclipse, git, hadoop, Web Scraping) statistics (regression, PCA, time series, trading) and more...

If you got this far, why not subscribe for updates from the site? Choose your flavor: e-mail, twitter, RSS, or facebook...

Tags: , , , , ,

Comments are closed.


Mango solutions

RStudio homepage

Zero Inflated Models and Generalized Linear Mixed Models with R

Quantide: statistical consulting and training


CRC R books series

Contact us if you wish to help support R-bloggers, and place your banner here.

Never miss an update!
Subscribe to R-bloggers to receive
e-mails with the latest R posts.
(You will not see this message again.)

Click here to close (This popup will not appear again)