Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.
Let us get back on the Titanic dataset,
loc_fichier = "http://freakonometrics.free.fr/titanic.RData"
download.file(loc_fichier, "titanic.RData")
load("titanic.RData")
base = base[!is.na(base$Age),]
On consider two variables, the age \(x\) (the continuous one) and the survivor indicator \(y\) (the qualitative one)
X = base$Age Y = base$Survived
It looks like the age might be a valid explanatory variable in the logistic regression,
summary(glm(Survived~Age,data=base,family=binomial))
Coefficients:
Estimate Std. Error z value Pr(>|z|)
(Intercept) -0.05672 0.17358 -0.327 0.7438
Age -0.01096 0.00533 -2.057 0.0397 *
---
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1
(Dispersion parameter for binomial family taken to be 1)
Null deviance: 964.52 on 713 degrees of freedom
Residual deviance: 960.23 on 712 degrees of freedom
AIC: 964.23
The significance test here has a \(p\)-value just below \(4%\). Actually, one can relate it with the value of the deviance (the null deviance and the residual deviance). Recall that\(D=2\big(\log\mathcal{L}(\boldsymbol{y})-\log\mathcal{L}(\widehat{\boldsymbol{\mu}})\big)\)while\(D_0=2\big(\log\mathcal{L}(\boldsymbol{y})-\log\mathcal{L}(\overline{y})\big)\)Under the assumption that \(x\) is worthless, \(D_0-D\) tends to a \(\chi^2\) distribution with 1 degree of freedom. And we can compute the \(p\)-value dof that likelihood ratio test,
1-pchisq(964.52-960.23,1) [1] 0.03833717
(which is consistent with a Gaussian test). But if we consider a nonlinear transformation
summary(glm(Survived~bs(Age),data=base,family=binomial))
Coefficients:
Estimate Std. Error z value Pr(>|z|)
(Intercept) 0.8648 0.3460 2.500 0.012433 *
bs(Age)1 -3.6772 1.0458 -3.516 0.000438 ***
bs(Age)2 1.7430 1.1068 1.575 0.115299
bs(Age)3 -3.9251 1.4544 -2.699 0.006961 **
---
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1
(Dispersion parameter for binomial family taken to be 1)
Null deviance: 964.52 on 713 degrees of freedom
Residual deviance: 948.69 on 710 degrees of freedom
which seems to be “more significant”
1-pchisq(964.52-948.69,3) [1] 0.001228712
So it looks like the variable \(x\) is interesting here.
To visualize the non-null correlation, one can consider the condition distribution of \(x\) given \(y=1\), and compare it with the condition distribution of \(x\) given \(y=0\),
ks.test(X[Y==0],X[Y==1]) Two-sample Kolmogorov-Smirnov test data: X[Y == 0] and X[Y == 1] D = 0.088777, p-value = 0.1324 alternative hypothesis: two-sided
i.e. with a \(p\)-value above \(10%\), the two distributions are not significatly different.
F0 = function(x) mean(X[Y==0]<=x) F1 = function(x) mean(X[Y==1]<=x) vx = seq(0,80,by=.1) vy0 = Vectorize(F0)(vx) vy1 = Vectorize(F1)(vx) plot(vx,vy0,col="red",type="s") lines(vx,vy1,col="blue",type="s")
(we can also look at the density, but it looks like that there is not much to see)
An alternative is discretize variable \(x\) and to use Pearson’s independence test,
k=5
LV = quantile(X,(0:k)/k)
LV[1] = 0
Xc = cut(X,LV)
table(Xc,Y)
Y
Xc 0 1
(0,19] 85 79
(19,25] 92 45
(25,31.8] 77 50
(31.8,41] 81 63
(41,80] 89 53
chisq.test(table(Xc,Y))
Pearson's Chi-squared test
data: table(Xc, Y)
X-squared = 8.6155, df = 4, p-value = 0.07146
The \(p\)-value is here \(7%\), with five categories for the age. And actually, we can compare the \(p\)-value
pvalue = function(k=5){
LV = quantile(X,(0:k)/k)
LV[1] = 0
Xc = cut(X,LV)
chisq.test(table(Xc,Y))$p.value}
vk = 2:20
vp = Vectorize(pvalue)(vk)
plot(vk,vp,type="l")
abline(h=.05,col="red",lty=2)
which gives a \(p\)-value close to \(5\)%, as soon as we have enough categories. In the slides of the course (STT5100), I claim that actually, the age is an important variable when trying to predict if a passenger survived. Test mentioned here are not as conclusive, nevertheless…
R-bloggers.com offers daily e-mail updates about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/data-science job.
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.
