You are on page 1of 41

The Simple Regression Model

y = 0 + 1x + u

Economics 20 - Prof. Anderson

Some Terminology
In the simple linear regression model, where y = 0 + 1x + u, we typically refer to y as the

Dependent Variable, or Left-Hand Side Variable, or Explained Variable, or Regressand

Economics 20 - Prof. Anderson

Some Terminology, cont.


In the simple linear regression of y on x, we typically refer to x as the

Independent Variable, or Right-Hand Side Variable, or Explanatory Variable, or Regressor, or Covariate, or Control Variables
Economics 20 - Prof. Anderson 3

A Simple Assumption
The average value of u, the error term, in the population is 0. That is, E(u) = 0 This is not a restrictive assumption, since we can always use 0 to normalize E(u) to 0
Economics 20 - Prof. Anderson 4

Zero Conditional Mean


We need to make a crucial assumption about how u and x are related We want it to be the case that knowing something about x does not give us any information about u, so that they are completely unrelated. That is, that E(u|x) = E(u) = 0, which implies E(y|x) = 0 + 1x
Economics 20 - Prof. Anderson 5

E(y|x) as a linear function of x, where for any x the distribution of y is centered about E(y|x)
y
f(y)

.
x1 x2
Economics 20 - Prof. Anderson

. E(y|x) = + x
0 1

Ordinary Least Squares


Basic idea of regression is to estimate the population parameters from a sample Let {(xi,yi): i=1, ,n} denote a random sample of size n from the population For each observation in this sample, it will be the case that yi = 0 + 1xi + ui
Economics 20 - Prof. Anderson 7

Population regression line, sample data points and the associated error terms
y y4 E(y|x) = 0 + 1x . u4 {

y3 y2

u2 {.

.} u3

y1

} u1

x1

x2

x3

x4

x
8

Economics 20 - Prof. Anderson

Deriving OLS Estimates


To derive the OLS estimates we need to realize that our main assumption of E(u|x) = E(u) = 0 also implies that Cov(x,u) = E(xu) = 0 Why? Remember from basic probability that Cov(X,Y) = E(XY) E(X)E(Y)
Economics 20 - Prof. Anderson 9

Deriving OLS continued


We can write our 2 restrictions just in terms of x, y, 0 and 1 , since u = y 0 1x E(y 0 1x) = 0 E[x(y 0 1x)] = 0 These are called moment restrictions Economics 20 - Prof. Anderson
10

Deriving OLS using M.O.M.


The method of moments approach to estimation implies imposing the population moment restrictions on the sample moments What does this mean? Recall that for E(X), the mean of a population distribution, a sample estimator of E(X) is simply the arithmetic mean of the sample
Economics 20 - Prof. Anderson 11

More Derivation of OLS


We want to choose values of the parameters that will ensure that the sample versions of our moment restrictions are true The sample versions are as follows:

n n

( y
n i= 1 n i= 1

0 1 xi = 0

xi yi 0 1 xi = 0
Economics 20 - Prof. Anderson 12

More Derivation of OLS


Given the definition of a sample mean, and properties of summation, we can rewrite the first condition as follows

y = 0 + 1 x , or 0 = y 1 x
Economics 20 - Prof. Anderson 13

More Derivation of OLS


xi yi y 1 x 1 xi = 0
i =1 n n

( (

xi ( yi y ) = 1 xi ( xi x )
i =1 n i =1 2 ( xi x )( yi y ) = 1 ( xi x ) i =1 i =1
Economics 20 - Prof. Anderson 14

So the OLS estimated slope is

1 =

( x x )( y
i =1 i n i =1 i n

y)
2

( x x)
i =1

provided that ( xi x ) > 0


2
Economics 20 - Prof. Anderson 15

Summary of OLS slope estimate


The slope estimate is the sample covariance between x and y divided by the sample variance of x If x and y are positively correlated, the slope will be positive If x and y are negatively correlated, the slope will be negative Only need x to vary in our sample
Economics 20 - Prof. Anderson 16

More OLS
Intuitively, OLS is fitting a line through the sample points such that the sum of squared residuals is as small as possible, hence the term least squares The residual, , is an estimate of the error term, u, and is the difference between the fitted line (sample regression function) and the sample point
Economics 20 - Prof. Anderson 17

Sample regression line, sample data points and the associated estimated error terms
y y4 4 {

.
y = 0 + 1 x

y3 y2

2 {.

.} 3

y1

} 1 .
x1 x2 x3 x4 x
18 Economics 20 - Prof. Anderson

Alternate approach to derivation


Given the intuitive idea of fitting a line, we can set up a formal minimization problem That is, we want to choose our parameters such that we minimize the following:

i ) = yi 0 1 xi (u
2 i =1 i =1
Economics 20 - Prof. Anderson

19

Alternate approach, continued


If one uses calculus to solve the minimization problem for the two parameters you obtain the following first order conditions, which are the same as we obtained before, multiplied by n

(y
n i =1 n i

1 xi = 0 0

xi yi 0 1 xi = 0
i =1
Economics 20 - Prof. Anderson 20

Algebraic Properties of OLS


The sum of the OLS residuals is zero Thus, the sample average of the OLS residuals is zero as well The sample covariance between the regressors and the OLS residuals is zero The OLS regression line always goes through the mean of the sample
Economics 20 - Prof. Anderson 21

Algebraic Properties (precise)


ui = 0 and thus,
i =1 n n

u
i =1

=0

x u
i =1 i

=0

y = 0 + 1 x
Economics 20 - Prof. Anderson 22

More terminology
We can think of each observatio n as being made up of an explained part, and an unexplaine d part, yi = yi + ui We then define the following :

Then SST = SSE + SSR

( y y ) is the total sum of squares (SST) ( y y ) is the explained sum of squares (SSE) u is the residual sum of squares (SSR)
2 2 i i 2 i
Economics 20 - Prof. Anderson 23

Proof that SST = SSE + SSR


( y y ) = [( y y ) + ( y y )] = [u + ( y y ) ] = u + 2 u ( y y ) + ( y y ) = SSR + 2 u ( y y ) + SSE and we know that u ( y y ) = 0
2 i i i i 2 i i 2 i i i i i i i i
Economics 20 - Prof. Anderson

24

Goodness-of-Fit
How do we think about how well our sample regression line fits our sample data? Can compute the fraction of the total sum of squares (SST) that is explained by the model, call this the R-squared of regression R2 = SSE/SST = 1 SSR/SST
Economics 20 - Prof. Anderson 25

Using Stata for OLS regressions


Now that weve derived the formula for calculating the OLS estimates of our parameters, youll be happy to know you dont have to compute them by hand Regressions in Stata are very simple, to run the regression of y on x, just type reg y x
Economics 20 - Prof. Anderson 26

Unbiasedness of OLS
Assume the population model is linear in parameters as y = 0 + 1x + u Assume we can use a random sample of size n, {(xi, yi): i=1, 2, , n}, from the population model. Thus we can write the sample model yi = 0 + 1xi + ui Assume E(u|x) = 0 and thus E(ui|xi) = 0 Assume there is variation in the xi
Economics 20 - Prof. Anderson 27

Unbiasedness of OLS (cont)


In order to think about unbiasedness, we need to rewrite our estimator in terms of the population parameter Start with a simple rewrite of the formula as

1
2 x

( x x) y =
i

2 x

, where

s ( xi x )

Economics 20 - Prof. Anderson

28

Unbiasedness of OLS (cont)

( x x ) y = ( x x )( + x ( x x ) + ( x x ) x + ( x x )u = ( x x ) + ( x x )x + ( x x )u
i i i i 0 0 i 1 i i i 0 i 1 i i i i
Economics 20 - Prof. Anderson

1 i

+ ui ) =

29

Unbiasedness of OLS (cont)

( x x ) = 0, ( x x)x = ( x x)
i i i i 2 1 x

so, the numerator can be rewritten as

s + ( xi x )ui , and thus


1 = 1

( x x )u +
i

2 x

Economics 20 - Prof. Anderson

30

Unbiasedness of OLS (cont)


let d i = ( xi x ) , so that = + 1 2 d u , then i i i 1 sx = + 1 2 d E( u ) = E 1 i 1 i 1 sx
Economics 20 - Prof. Anderson 31

( )

Unbiasedness Summary
The OLS estimates of 1 and 0 are unbiased Proof of unbiasedness depends on our 4 assumptions if any assumption fails, then OLS is not necessarily unbiased Remember unbiasedness is a description of the estimator in a given sample we may be near or far from the true parameter
Economics 20 - Prof. Anderson 32

Variance of the OLS Estimators


Now we know that the sampling distribution of our estimate is centered around the true parameter Want to think about how spread out this distribution is Much easier to think about this variance under an additional assumption, so Assume Var(u|x) = 2 (Homoskedasticity)
Economics 20 - Prof. Anderson 33

Variance of OLS (cont)


Var(u|x) = E(u2|x)-[E(u|x)]2 E(u|x) = 0, so 2 = E(u2|x) = E(u2) = Var(u) Thus 2 is also the unconditional variance, called the error variance , the square root of the error variance is called the standard deviation of the error Can say: E(y|x)=0 + 1x and Var(y|x) = 2
Economics 20 - Prof. Anderson 34

Homoskedastic Case
y
f(y|x)

.
x1 x2
Economics 20 - Prof. Anderson

. E(y|x) = + x
0 1

35

Heteroskedastic Case
f(y|x)
y

.
x1 x2 x3

.
Economics 20 - Prof. Anderson

E(y|x) = 0 + 1x

x
36

Variance of OLS (cont)


= Var + 1 2 d u = 1 Var 1 i i sx 1 1 2 Var ( d i ui ) = 2 sx sx 1 = 2 sx
2 2 2 i 2 2 2 2

( )

d i2Var ( ui )
2

1 d = sx2
2

d i2 =

1 2 2 = Var 1 sx = 2 2 sx sx
Economics 20 - Prof. Anderson

( )
37

Variance of OLS Summary


The larger the error variance, 2, the larger the variance of the slope estimate The larger the variability in the xi, the smaller the variance of the slope estimate As a result, a larger sample size should decrease the variance of the slope estimate Problem that the error variance is unknown
Economics 20 - Prof. Anderson 38

Estimating the Error Variance


We dont know what the error variance, 2, is, because we dont observe the errors, ui What we observe are the residuals, i We can use the residuals to form an estimate of the error variance
Economics 20 - Prof. Anderson 39

Error Variance Estimate (cont)


ui = yi 0 1 xi = ( 0 + 1 xi + ui ) 0 1 xi =u
i

) (

Then, an unbiased estimator of 2 is 1 2 = i2 = SSR / ( n 2) u ( n 2)


Economics 20 - Prof. Anderson 40

Error Variance Estimate (cont)


= 2 = Standard error of the regression recall that sd = sx if we substitute for then we have the standard error of ,

( )

se 1 = / ( xi x )

( )

Economics 20 - Prof. Anderson

41

You might also like