# Generalized linear models
# Logistic regression on Titanic dataset
Logistic regression is a particular case of the generalized linear model, used to model dichotomous outcomes (probit and complementary loglog models are closely related).
The name comes from the link function used, the logit or logodds function. The inverse function of the logit is called the logistic function and is given by:
This function takes a value between ]Inf;+Inf[ and returns a value between 0 and 1; i.e the logistic function takes a linear predictor and returns a probability.
Logistic regression can be performed using the glm
function with the option family = binomial
(shortcut for family = binomial(link="logit")
; the logit being the default link function for the binomial family).
In this example, we try to predict the fate of the passengers aboard the RMS Titanic.
Read the data:
url < "http://biostat.mc.vanderbilt.edu/wiki/pub/Main/DataSets/titanic.txt"
titanic < read.csv(file = url, stringsAsFactors = FALSE)
Clean the missing values:
In that case, we replace the missing values by an approximation, the average.
titanic$age[is.na(titanic$age)] < mean(titanic$age, na.rm = TRUE)
Train the model:
titanic.train < glm(survived ~ pclass + sex + age,
family = binomial, data = titanic)
Summary of the model:
summary(titanic.train)
The output:
Call:
glm(formula = survived ~ pclass + sex + age, family = binomial, data = titanic)
Deviance Residuals:
Min 1Q Median 3Q Max
2.6452 0.6641 0.3679 0.6123 2.5615
Coefficients:
Estimate Std. Error z value Pr(>z)
(Intercept) 3.552261 0.342188 10.381 < 2e16 ***
pclass2nd 1.170777 0.211559 5.534 3.13e08 ***
pclass3rd 2.430672 0.195157 12.455 < 2e16 ***
sexmale 2.463377 0.154587 15.935 < 2e16 ***
age 0.042235 0.007415 5.696 1.23e08 ***

Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1
(Dispersion parameter for binomial family taken to be 1)
Null deviance: 1686.8 on 1312 degrees of freedom
Residual deviance: 1165.7 on 1308 degrees of freedom
AIC: 1175.7
Number of Fisher Scoring iterations: 5

 The qualitative variables are "dummified". A modality is considered as the reference. The reference modality can be change with `I` in the formula.
 All four predictors are statistically significant at a 0.1 % level.
 The logistic regression coefficients give the change in the log odds of the outcome for a one unit increase in the predictor variable.
 To see the **odds ratio** (multiplicative change in the odds of survival per unit increase in a predictor variable), exponentiate the parameter.
 To see the confidence interval (CI) of the parameter, use `confint`.
 When comparing models fitted by maximum likelihood to the same data, the smaller the AIC, the better the fit.
 One measure of model fit is the significance of the overall model. This test asks whether the model with predictors fits significantly better than a model with just an intercept (i.e., a null model).
Below the table of coefficients are fit indices, including the null and deviance residuals and the Akaike Information Criterion (AIC), which can be used for comparing model performance.
Example of odds ratios:
exp(coef(titanic.train)[3])
pclass3rd
0.08797765
With this model, compared to the first class, the 3rd class passengers have about a tenth of the odds of survival.
Example of confidence interval for the parameters:
confint(titanic.train)
Waiting for profiling to be done...
2.5 % 97.5 %
(Intercept) 2.89486872 4.23734280
pclass2nd 1.58986065 0.75987230
pclass3rd 2.81987935 2.05419500
sexmale 2.77180962 2.16528316
age 0.05695894 0.02786211
Exemple of calculating the significance of the overall model:
The test statistic is distributed chisquared with degrees of freedom equal to the differences in degrees of freedom between the current and the null model (i.e., the number of predictor variables in the model).
with(titanic.train, pchisq(null.deviance  deviance, df.null  df.residual
, lower.tail = FALSE))
[1] 1.892539e111
The pvalue is near 0, showing a strongly significant model.