Professional Documents
Culture Documents
y 0 1 x1 2 x2 u (3.40)
~ ~
y 0 1 x1
~ ~
1 1 2 (3.43)
-where Bhats come from regressing y on ALL xs and deltatilde comes
from regressing x2 on x1
-since deltatilde depends on independent variables, it is considered fixed
-we also know from Theorem 3.1 that Bhats are unbiased estimators,
therefore:
~
~
E ( 1 ) 1 2 (3.45)
Bias ( 1 ) E ( 1 ) 1 2 (3.46)
-this bias is often called OMITTED VARIABLE BIAS
-From this equation, B1tilde is unbiased in two cases:
1) B2=0; x2 has no impact on y in the true model
2) deltatilde=0
3.3 Deltatilde=0
-deltatilde is equal to the covariance of x 1 and
x2 over the variance of x1, all in the sample
-deltatilde is equal to zero only if x 1 and x2 are
uncorrelated
-therefore if they are uncorrelated, B 1hat is
unbiased
-it is also unbiased if we can show that:
E ( x2 | x1 ) E ( x2 )
B2hat>0
Positive Bias
Negative Bias
B2hat<0
Negative Bias
Positive Bias
3.3 Example
(ie)
(ie)
3.3 Example
Pasta 5.3 0.4Experience
(ie)
E ( 1 ) 1
3.3
General
Omitted
Bias
Deriving the direction of omitted variable bias with more independent
variables is more difficult
-Note that correlation between any explanatory variable and the error causes
ALL OLS estimates to be biased.
-Consider the true and estimated models:
y 0 1x1 2 x 2 3 x 3 u
~ ~
~
~
y x x
0
1 1
(3.49)
(3.50)
3.3
General
Omitted
Bias
Since our x values can be pairwise correlated, it is hard to derive the
bias for our OLS estimates
-If we assume that x1 and x2 are uncorrelated, we can analyze B1tildes
bias without x2 having an effect, similar to our 2 variable regression:
( xi1 x1 ) xi 3
~
E ( 1 ) 1 3
2
( xi1 x1 )
With this formula similar to (3.45), the previous table can be used to determine bias
-Note that much uncorrelation is needed to determine bias
Assumption MLR.5
(Homoskedasticity)
The error u has the same variance
given any values of the
explanatory variables. In other
words,
Var (u | x1 , x2 ,..., xk )
Assumption MLR.5
Notes
-MLR. 5 assumes that the
variance of the error term, u, is
the SAME for ANY combination of explanatory variables
-If ANY explanatory variable affects the errors variance,
HETEROSKEDASTICITY is present
-The above five assumptions are called the GAUSSMARKOV ASSUMPTIONS
-As listed above, they apply only to cross-sectional data
with random sampling
-time series and panel data analysis require more
complicated, related assumptions
Assumption MLR.5
Notes
If we let X represent
all x variables,
combining assumptions 1 through 4 give
us: E (y | X) 0 1x1 2 x 2 ... k x k
Or as an example:
E (U | X) 0 1 I 2 Time 3Love
Var ( y | X )
Or for example:
Var (U | X )
Theorem 3.2
(Sampling Variances of
the OLS Slope Estimators)
Under assumptions MLR. 1 through MRL. 5,
conditional on the sample values of the
independent variables,
2
Var ( j )
(3.51)
2
SST j (1 R j )
ij