# Simulating Endogeneity

March 9, 2015
By

(This article was first published on Stable Markets » R, and kindly contributed to R-bloggers)

Introduction The topic in this post is endogeneity, which can severely bias regression estimates. I will specifically simulate endogeneity caused by an omitted variable. In future posts in this series, I’ll simulate other specification issues such as heteroskedasticity, multicollinearity, and collider bias.

The Data-Generating Process

Consider the data-generating process (DGP) of some outcome variable $Y$:

$Y = a+\beta x+c z + \epsilon_1$

$\epsilon_1 \sim N(0,\sigma^{2})$

For the simulation, I set parameter values for $a$, $\beta$, and $c$ and simulate positively correlated independent variables, $x$ and $z$ (N=500).

# simulation parameters
set.seed(144);
ss=500; trials=5000;
a=50; b=.5; c=.01;
d=25; h=.9;

# generate two independent variables
x=rnorm(n=ss,mean=1000,sd=50);
z=d+h*x+rnorm(ss,0,10)


The Simulation

The simulation will estimate the two models below. The first model is correct in the sense that it includes all terms in the actual DGP. However, the second model omits a variable that is present in the DGP. Instead, the variable is obsorbed into the error term $\epsilon_1$.

$(1)\thinspace Y = a+\beta x+c z + \epsilon_1$

$(2) \thinspace Y = a+\beta x + \epsilon_1$

This second model will yield a biased estimator of $\beta$. The variance will also be biased. This is because $x$ is endogenous, which is a fancy way of saying it is correlated with the error term, $\epsilon_1$. Since

sim=function(endog){
# assume normal error with constant variance to start
e=rnorm(n=ss,mean=0,sd=10)
y=a+b*x+c*z+e
# Select data generation process
if(endog==TRUE){ fit=lm(y~x) }else{ fit=lm(y~x+z)}
return(fit\$coefficients)
}

# run simulation - with and wihtout endogeneity
sim_results=t(replicate(trials,sim(endog=FALSE)))
sim_results_endog=t(replicate(trials,sim(endog=TRUE)))


Simulation Results This simulation yields two different sampling distributions for $\beta$. Note that I have set the true value to $\beta=.5$. When $z$ is not omitted, the simulation yields the green sampling distribution, centered around the true value. The average value across all simulations is 0.4998. When $z$ is omitted, the simulation yields the red sampling distribution, centered around 0.5895. It’s biased upward from the true value of .5 by .0895. Moreover, the variance of the biased sampling distribution is much smaller than the true variance around $\beta$. This compromises the ability to perform any meaningful inferences about the true parameter. Bias Analysis The bias in $\beta$ can be derived analytically. Consider that in model 1 (presented above), $x$ and $z$ are related by:

$(3)\thinspace z = d+hx+\epsilon_2$

Substituting $z$ in equation 1 with equation 3 and re-ordering:

$Y = a+\beta x+c (d+hx+\epsilon_2) + \epsilon_1$

$(4)\thinspace Y = (a+cd)+(\beta+ch) x + (\epsilon_1+c\epsilon_2)$

When omitting variable $z$, it is actually equation 4 that is estimated. It can be seen that $\beta$ is biased by the quantity $ch$. In this case, since $x$ and $z$ are positively correlated by construction and their slope coefficients are positive, the bias will be positive. According to the parameters of the simulation, the “true” bias should be  $ch=.09$. Here is the distribution of the bias, it is centered around .0895, very close to the true bias value. The derivation above also lets us determine the direction of bias from knowing the correlation of $x$ and $z$ as well as the sign of $c$ (the true partial effect of $z$ on $y$). If both are the same sign, then the estimate of $\beta$ will be biased upward. If the signs differ, then the estimate of $\beta$ will be biased downward. Conclusion The case above was pretty general, but has particular applications. For example, if we believe that an individual’s income is a function of years of education and year of work experience, then omitting one variable will bias the slope estimate of the other.

To leave a comment for the author, please follow the link and comment on their blog: Stable Markets » R.

R-bloggers.com offers daily e-mail updates about R news and tutorials on topics such as: Data science, Big Data, R jobs, visualization (ggplot2, Boxplots, maps, animation), programming (RStudio, Sweave, LaTeX, SQL, Eclipse, git, hadoop, Web Scraping) statistics (regression, PCA, time series, trading) and more...

If you got this far, why not subscribe for updates from the site? Choose your flavor: e-mail, twitter, RSS, or facebook...