# Introduction to Logistic Regression with R

**Data Perspective**, and kindly contributed to R-bloggers]. (You can report issue about the content on this page here)

Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.

Consider a scenario where we need to predict a medical condition of a patient (HBP) ,HAVE HIGH BP or NO HIGH BP, based on some observed symptoms – Age, weight, Issmoking, Systolic value, Diastolic value, RACE, etc.. In this scenario we have to build a model which takes the above mentioned symptoms as input values and HBP as response variable. Note that the response variable (HBP) is a value among a fixed set of classes, HAVE HIGH BP or NO HIGH BP.

### Logistic regression – a classification problem, not a prediction problem:

In my previous blog I told that we use linear regression for scenarios which involves prediction. But there is a check; the regression analysis cannot be applied in scenarios where the response variable is not continuous. In our case the response variable is not a continuous variable but a value among a fixed set of classes. We call such scenarios as Classification problem rather than prediction problem. In such scenarios where the response variables are more of qualitative nature rather than continuous nature, we have to apply more suitable models namely logistic regression for classification.

Assume we have a binary category output variable Y and a vector of p input variables X. Rather than modeling this response Y directly, logistic regression models the conditional probability, **Pr(Y = 1|X = x)** as a function of x, that Y belongs to a particular category.

Mathematically, logistic regression is expressed as:

**Estimating Coefficients – Maximum Likelihood function:**

The unknown parameters, β0/ β1, in the function are estimated by maximum likelihood method using available input training data. The Maximum likelihood function expresses the probability of the observed data as a function of the unknown parameters. The maximum likelihood estimators of these parameters are chosen to be those values that maximize this function. Thus, the estimators are those which agree most closely with the observed data.

For now we assume that solving the above equation can be used to estimate the unknown parameters.

In R, we glm() which takes training data as input and gives us the fitted model with estimated parameters as output, which we will see in the later section.

#### Making Predictions:

Once the coefficients have been estimated, it is a simple matter to compute the probability of response variable for any given input values by putting values of β0/ β1/X in the below equation.

*Note: we have predict() in R which takes fitted model, input parameters as input values to predict the response variables.*

#### Use case – Classify if a person may have HBP/not HBP :

Let us take a use case and implement logistic regression in R. Let us classify/predict if a person suffers with High Blood Pressure (HPB) given input predictors AGE, SEX, IsSmoking, Avg Systolic BP, Average diastolic BP, RACE, Body Weight, Height, etc. Access the data set from here. The dataset is NHANES III dataset. In order to implement logistic regression, we need to follow the below steps:

*data = read.csv(‘~nhanes.csv’)*

*str(data)*

*#removing SEQN,SDPPSU6,SDPSTRA6 from the dataset, from description of the datset,*

*#we concluded that it is not related to data.*

*data = data[,-c(1:3)]*

*#excluding HAR3 for now.*

*data = data[,-c(1,10)]*

*# removing entire row if there is any single missing column in the data.*

*#check for NA’s or missing values in the data*

*#removing all the na rows from data*

*data = na.omit(data)*

*# from the description we consider HBP as response variable.*

*names(data)*

*summary(data)*

*‘Creating dummy variables/design variables for DMARACER/HSSEX create number of dummy variables based on the levels -1, i.e if if there 3 levels create 3-1 levels of dummy variables x= data*

*for RACE create 2 dummy variables DMRACE1, DMRACE2 where White(1) – (0,0) black(2) (1,0),other (3) – (1,1)’*

*for(level in unique(data$DMARACER)){*

*data[paste(“RACE”, level, sep = “_”)] <- ifelse(data$DMARACER == level, 1, 0)*

*}*

*#removing race_1*

*data = data[,-12]*

*for(level in unique(data$SMOKE)){*

*data[paste(“SMOKE”, level, sep = “_”)] <- ifelse(data$SMOKE == level, 1, 0)*

*}*

*#removing SMOKE_1*

*data = data[,-14]*

*table(data$HSSEX,data$HBP)*

0 1 0 6576 1702 1 5817 1548

*Fem_HBP = (1702/6576+1702)) = 0.2056052*

*Mal_HBP = (1548/(1548+5817)) = 0.2101833*

*Modelling:*

*#remove HAR1,PEPMNK1R as SMOKE & HBP are created from these variables.*

*data = data[,-c(3,6,7,8,9,16)]*

*#create training & testing data*

*train_x = data*

*smp_size <- floor(0.75 * nrow(train_x))*

*train_ind <- sample(seq_len(nrow(train_x)), size = smp_size)*

*train <- train_x[train_ind, ]*

*test <- train_x[-train_ind, ]*

*glm2 = glm(train$HBP~train$HSAGEIR+train$HSSEX+train$BMPWTLBS+train$BMPHTIN+train$TCP+train$RACE_2+train$RACE_3,data=train,family=binomial)*

The image shows us the summary of the model. In my next post let us evaluate the logistic regression model and let us consider few other models and choose better model among them.

**leave a comment**for the author, please follow the link and comment on their blog:

**Data Perspective**.

R-bloggers.com offers

**daily e-mail updates**about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/data-science job.

Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.