Skip to main content
added 3 characters in body
Source Link
Richard Hardy
  • 71.5k
  • 14
  • 129
  • 292

In Introductory Econometrics: A Modern Approach, Wooldridge writes the following regarding the omitted variable bias and its effect on the variance of the OLS estimator (x1 and x2 are correlated):

enter image description here

This intutivleyintuitively makes sense given that by defintiondefinition of the omitted variable bias x1 and x2 are correlated. Thus including x2 into the regression should inflate the variance due to multicollinearity.

However when I run a simulation in R I consitentlyconsistently get the exact opposite of what Wooldridge suggest. Consider the data generating process:

x1 <- rnorm(10000) x2 <- rnorm(10000) + 0.2*x1 y <- 0.5 -2*x1 -2.5*x2 + rnorm(10000) summary(lm(y ~ x1 + x2)) summary(lm(y ~ x1)) 

No matter how many times I run this simulation the standard error of beta1 in the omitted variable case is always larger than in the unbaisedunbiased specification.

How is this possible?

In Introductory Econometrics: A Modern Approach, Wooldridge writes the following regarding the omitted variable bias and its effect on the variance of the OLS estimator (x1 and x2 are correlated):

enter image description here

This intutivley makes sense given that by defintion of the omitted variable bias x1 and x2 are correlated. Thus including x2 into the regression should inflate the variance due to multicollinearity.

However when I run a simulation in R I consitently get the exact opposite of what Wooldridge suggest. Consider the data generating process:

x1 <- rnorm(10000) x2 <- rnorm(10000) + 0.2*x1 y <- 0.5 -2*x1 -2.5*x2 + rnorm(10000) summary(lm(y ~ x1 + x2)) summary(lm(y ~ x1)) 

No matter how many times I run this simulation the standard error of beta1 in the omitted variable case is always larger than in the unbaised specification.

How is this possible?

In Introductory Econometrics: A Modern Approach, Wooldridge writes the following regarding the omitted variable bias and its effect on the variance of the OLS estimator (x1 and x2 are correlated):

enter image description here

This intuitively makes sense given that by definition of the omitted variable bias x1 and x2 are correlated. Thus including x2 into the regression should inflate the variance due to multicollinearity.

However when I run a simulation in R I consistently get the exact opposite of what Wooldridge suggest. Consider the data generating process:

x1 <- rnorm(10000) x2 <- rnorm(10000) + 0.2*x1 y <- 0.5 -2*x1 -2.5*x2 + rnorm(10000) summary(lm(y ~ x1 + x2)) summary(lm(y ~ x1)) 

No matter how many times I run this simulation the standard error of beta1 in the omitted variable case is always larger than in the unbiased specification.

How is this possible?

Tweeted twitter.com/StackStats/status/1099277856075595779
Source Link
Jonas C
  • 103
  • 6

Omitted Variable Bias & Multicollinearity: Why are the coefficient SEs smaller in the unbiased specification?

In Introductory Econometrics: A Modern Approach, Wooldridge writes the following regarding the omitted variable bias and its effect on the variance of the OLS estimator (x1 and x2 are correlated):

enter image description here

This intutivley makes sense given that by defintion of the omitted variable bias x1 and x2 are correlated. Thus including x2 into the regression should inflate the variance due to multicollinearity.

However when I run a simulation in R I consitently get the exact opposite of what Wooldridge suggest. Consider the data generating process:

x1 <- rnorm(10000) x2 <- rnorm(10000) + 0.2*x1 y <- 0.5 -2*x1 -2.5*x2 + rnorm(10000) summary(lm(y ~ x1 + x2)) summary(lm(y ~ x1)) 

No matter how many times I run this simulation the standard error of beta1 in the omitted variable case is always larger than in the unbaised specification.

How is this possible?