+ All Categories
Home > Documents > 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y = 0 + 1 x + u.

1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y = 0 + 1 x + u.

Date post: 21-Dec-2015
Category:
View: 216 times
Download: 0 times
Share this document with a friend
41
1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y = 0 + 1 x + u
Transcript
Page 1: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

1Prof. Dr. Rainer Stachuletz

The Simple Regression Model

y = 0 + 1x + u

Page 2: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

2Prof. Dr. Rainer Stachuletz

Some Terminology

In the simple linear regression model, where y = 0 + 1x + u, we typically refer to y as the Dependent Variable, or Left-Hand Side Variable, or Explained Variable, or Regressand

Page 3: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

3Prof. Dr. Rainer Stachuletz

Some Terminology, cont.

In the simple linear regression of y on x, we typically refer to x as the Independent Variable, or Right-Hand Side Variable, or Explanatory Variable, or Regressor, or Covariate, or Control Variables

Page 4: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

4Prof. Dr. Rainer Stachuletz

A Simple Assumption

The average value of u, the error term, in the population is 0. That is,

E(u) = 0

This is not a restrictive assumption, since we can always use 0 to normalize E(u) to 0

Page 5: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

5Prof. Dr. Rainer Stachuletz

Zero Conditional Mean

We need to make a crucial assumption about how u and x are related We want it to be the case that knowing something about x does not give us any information about u, so that they are completely unrelated. That is, that E(u|x) = E(u) = 0, which implies

E(y|x) = 0 + 1x

Page 6: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

6Prof. Dr. Rainer Stachuletz

..

x1 x2

E(y|x) as a linear function of x, where for any x the distribution of y is centered about E(y|x)

E(y|x) = 0 + 1x

y

f(y)

Page 7: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

7Prof. Dr. Rainer Stachuletz

Ordinary Least Squares

Basic idea of regression is to estimate the population parameters from a sample

Let {(xi,yi): i=1, …,n} denote a random sample of size n from the population

For each observation in this sample, it will be the case that

yi = 0 + 1xi + ui

Page 8: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

8Prof. Dr. Rainer Stachuletz

.

..

.

y4

y1

y2

y3

x1 x2 x3 x4

}

}

{

{

u1

u2

u3

u4

x

y

Population regression line, sample data pointsand the associated error terms

E(y|x) = 0 + 1x

Page 9: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

9Prof. Dr. Rainer Stachuletz

Deriving OLS Estimates

To derive the OLS estimates we need to realize that our main assumption of E(u|x) = E(u) = 0 also implies that

Cov(x,u) = E(xu) = 0

Why? Remember from basic probability that Cov(X,Y) = E(XY) – E(X)E(Y)

Page 10: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

10Prof. Dr. Rainer Stachuletz

Deriving OLS continued

We can write our 2 restrictions just in terms of x, y, 0 and , since u = y – 0 – 1x

E(y – 0 – 1x) = 0

E[x(y – 0 – 1x)] = 0

These are called moment restrictions

Page 11: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

11Prof. Dr. Rainer Stachuletz

Deriving OLS using M.O.M.

The method of moments approach to estimation implies imposing the population moment restrictions on the sample moments

What does this mean? Recall that for E(X), the mean of a population distribution, a sample estimator of E(X) is simply the arithmetic mean of the sample

Page 12: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

12Prof. Dr. Rainer Stachuletz

More Derivation of OLS

We want to choose values of the parameters that will ensure that the sample versions of our moment restrictions are true

The sample versions are as follows:

0ˆˆ

0ˆˆ

110

1

110

1

n

iiii

n

iii

xyxn

xyn

Page 13: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

13Prof. Dr. Rainer Stachuletz

More Derivation of OLS

Given the definition of a sample mean, and properties of summation, we can rewrite the first condition as follows

xy

xy

10

10

ˆˆ

or

,ˆˆ

Page 14: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

14Prof. Dr. Rainer Stachuletz

More Derivation of OLS

n

iii

n

ii

n

iii

n

iii

n

iiii

xxyyxx

xxxyyx

xxyyx

1

21

1

11

1

111

ˆ

ˆ

0ˆˆ

Page 15: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

15Prof. Dr. Rainer Stachuletz

So the OLS estimated slope is

0 that provided

ˆ

1

2

1

2

11

n

ii

n

ii

n

iii

xx

xx

yyxx

Page 16: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

16Prof. Dr. Rainer Stachuletz

Summary of OLS slope estimate

The slope estimate is the sample covariance between x and y divided by the sample variance of x If x and y are positively correlated, the slope will be positive If x and y are negatively correlated, the slope will be negative Only need x to vary in our sample

Page 17: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

17Prof. Dr. Rainer Stachuletz

More OLS

Intuitively, OLS is fitting a line through the sample points such that the sum of squared residuals is as small as possible, hence the term least squares

The residual, û, is an estimate of the error term, u, and is the difference between the fitted line (sample regression function) and the sample point

Page 18: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

18Prof. Dr. Rainer Stachuletz

.

..

.

y4

y1

y2

y3

x1 x2 x3 x4

}

}

{

{

û1

û2

û3

û4

x

y

Sample regression line, sample data pointsand the associated estimated error terms

xy 10ˆˆˆ

Page 19: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

19Prof. Dr. Rainer Stachuletz

Alternate approach to derivation

Given the intuitive idea of fitting a line, we can set up a formal minimization problem

That is, we want to choose our parameters such that we minimize the following:

n

iii

n

ii xyu

1

2

101

2 ˆˆˆ

Page 20: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

20Prof. Dr. Rainer Stachuletz

Alternate approach, continued

If one uses calculus to solve the minimization problem for the two parameters you obtain the following first order conditions, which are the same as we obtained before, multiplied by n

0ˆˆ

0ˆˆ

110

110

n

iiii

n

iii

xyx

xy

Page 21: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

21Prof. Dr. Rainer Stachuletz

Algebraic Properties of OLS

The sum of the OLS residuals is zero

Thus, the sample average of the OLS residuals is zero as well

The sample covariance between the regressors and the OLS residuals is zero

The OLS regression line always goes through the mean of the sample

Page 22: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

22Prof. Dr. Rainer Stachuletz

Algebraic Properties (precise)

xy

ux

n

uu

n

iii

n

iin

ii

10

1

1

1

ˆˆ

0

ˆ

thus,and 0ˆ

Page 23: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

23Prof. Dr. Rainer Stachuletz

More terminology

SSR SSE SSTThen

(SSR) squares of sum residual theis ˆ

(SSE) squares of sum explained theis ˆ

(SST) squares of sum total theis

:following thedefine then Weˆˆ

part, dunexplainean and part, explainedan of up

made being asn observatioeach ofcan think We

2

2

2

i

i

i

iii

u

yy

yy

uyy

Page 24: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

24Prof. Dr. Rainer Stachuletz

Proof that SST = SSE + SSR

0 ˆˆ that know weand

SSE ˆˆ2 SSR

ˆˆˆ2ˆ

ˆˆ

ˆˆ

22

2

22

yyu

yyu

yyyyuu

yyu

yyyyyy

ii

ii

iiii

ii

iiii

Page 25: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

25Prof. Dr. Rainer Stachuletz

Goodness-of-Fit

How do we think about how well our sample regression line fits our sample data?

Can compute the fraction of the total sum of squares (SST) that is explained by the model, call this the R-squared of regression

R2 = SSE/SST = 1 – SSR/SST

Page 26: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

26Prof. Dr. Rainer Stachuletz

Using Stata for OLS regressions

Now that we’ve derived the formula for calculating the OLS estimates of our parameters, you’ll be happy to know you don’t have to compute them by hand

Regressions in Stata are very simple, to run the regression of y on x, just type

reg y x

Page 27: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

27Prof. Dr. Rainer Stachuletz

Unbiasedness of OLS

Assume the population model is linear in parameters as y = 0 + 1x + u Assume we can use a random sample of size n, {(xi, yi): i=1, 2, …, n}, from the population model. Thus we can write the sample model yi = 0 + 1xi + ui

Assume E(u|x) = 0 and thus E(ui|xi) = 0

Assume there is variation in the xi

Page 28: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

28Prof. Dr. Rainer Stachuletz

Unbiasedness of OLS (cont)

In order to think about unbiasedness, we need to rewrite our estimator in terms of the population parameter

Start with a simple rewrite of the formula as

22

21 where,ˆ

xxs

s

yxx

ix

x

ii

Page 29: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

29Prof. Dr. Rainer Stachuletz

Unbiasedness of OLS (cont)

ii

iii

ii

iii

iiiii

uxx

xxxxx

uxx

xxxxx

uxxxyxx

10

10

10

Page 30: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

30Prof. Dr. Rainer Stachuletz

Unbiasedness of OLS (cont)

211

21

2

ˆ

thusand ,

asrewritten becan numerator the,so

,0

x

ii

iix

iii

i

s

uxx

uxxs

xxxxx

xx

Page 31: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

31Prof. Dr. Rainer Stachuletz

Unbiasedness of OLS (cont)

1211

21

then,1ˆ

thatso ,let

iix

iix

i

ii

uEds

E

uds

xxd

Page 32: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

32Prof. Dr. Rainer Stachuletz

Unbiasedness Summary

The OLS estimates of 1 and 0 are unbiased Proof of unbiasedness depends on our 4 assumptions – if any assumption fails, then OLS is not necessarily unbiased Remember unbiasedness is a description of the estimator – in a given sample we may be “near” or “far” from the true parameter

Page 33: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

33Prof. Dr. Rainer Stachuletz

Variance of the OLS Estimators

Now we know that the sampling distribution of our estimate is centered around the true parameter Want to think about how spread out this distribution is Much easier to think about this variance under an additional assumption, soAssume Var(u|x) = 2 (Homoskedasticity)

Page 34: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

34Prof. Dr. Rainer Stachuletz

Variance of OLS (cont)

Var(u|x) = E(u2|x)-[E(u|x)]2

E(u|x) = 0, so 2 = E(u2|x) = E(u2) = Var(u)

Thus 2 is also the unconditional variance, called the error variance

, the square root of the error variance is called the standard deviation of the error

Can say: E(y|x)=0 + 1x and Var(y|x) = 2

Page 35: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

35Prof. Dr. Rainer Stachuletz

..

x1 x2

Homoskedastic Case

E(y|x) = 0 + 1x

y

f(y|x)

Page 36: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

36Prof. Dr. Rainer Stachuletz

.

x x1 x2

yf(y|x)

Heteroskedastic Case

x3

..

E(y|x) = 0 + 1x

Page 37: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

37Prof. Dr. Rainer Stachuletz

Variance of OLS (cont)

12

222

22

22

2222

2

2

22

2

2

2

211

ˆ1

11

11

Vars

ss

ds

ds

uVards

udVars

uds

VarVar

xx

x

ix

ix

iix

iix

iix

Page 38: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

38Prof. Dr. Rainer Stachuletz

Variance of OLS Summary

The larger the error variance, 2, the larger the variance of the slope estimate

The larger the variability in the xi, the smaller the variance of the slope estimate

As a result, a larger sample size should decrease the variance of the slope estimate

Problem that the error variance is unknown

Page 39: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

39Prof. Dr. Rainer Stachuletz

Estimating the Error Variance

We don’t know what the error variance, 2, is, because we don’t observe the errors, ui

What we observe are the residuals, ûi

We can use the residuals to form an estimate of the error variance

Page 40: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

40Prof. Dr. Rainer Stachuletz

Error Variance Estimate (cont)

2/ˆ2

is ofestimator unbiasedan Then,

ˆˆ

ˆˆ

ˆˆˆ

22

2

1100

1010

10

nSSRun

u

xux

xyu

i

i

iii

iii

Page 41: 1 Prof. Dr. Rainer Stachuletz The Simple Regression Model y =  0 +  1 x + u.

41Prof. Dr. Rainer Stachuletz

Error Variance Estimate (cont)

21

21

1

2

/ˆˆse

, ˆ oferror standard the

have then wefor ˆ substitute weif

ˆsd that recall

regression theoferror Standardˆˆ

xx

s

i

x


Recommended