ch04ch05lectr06

89
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 6: Interpreting Regression Results Logarithms (Chapter 4.5) Standard Errors (Chapter 5)

Upload: vrajesh

Post on 25-Jan-2016

212 views

Category:

Documents


0 download

DESCRIPTION

lec 6

TRANSCRIPT

Page 1: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved.

Lecture 6:Interpreting Regression Results

Logarithms (Chapter 4.5)

Standard Errors(Chapter 5)

Page 2: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-2

Agenda for Today

• Review

• Logarithms in Econometrics (Chapter 4.5)

• Residuals (Chapter 5.1)

• Estimating an Estimator’s Variance (Chapter 5.2)

• Confidence Intervals (Chapter 5.4)

Page 3: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-3

Review

• As a starting place for analysis, we need to write down all our assumptions about the way the underlying process works, and about how that process led to our data.

• These assumptions are called the “Data Generating Process.”

• Then we can derive estimators that have good properties for the Data Generating Process we have assumed.

Page 4: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-4

Review: The Gauss–Markov DGP

• Y = X +• E(i ) = 0

• Var(i ) = 2

• Cov(i ,j ) = 0, for i ≠ j

• X ’s fixed across samples (so we can treat them like constants).

• We want to estimate

Page 5: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-5

Review

2

1

1

1

( ) 0

( ) ( , ) 0,

ˆ

ˆ( )

1.

i i i i

i i j

n

i ii

n

i ii

n

i ii

Y X E

Var Cov i j

X

wY

E w X

w X

for

's fixed across samples

A linear estimator is unbiased if

Page 6: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-6

Review (cont.)

A linear estimator is unbiased if wiX

ii1

n

1.

Many linear estimators will be unbiased.

How do I pick the "best" linear

unbiased estimator (BLUE)?

Page 7: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-7

Review: BLUE Estimators

A Strategy for Finding the Best Linear Unbiased Estimator:

1. Start with linear estimators: wiYi

– Impose the unbiasedness condition wiXi =1

– Calculate the variance of a linear estimator: Var(wiYi) =2wi

2

1. Use calculus to find the wi that give the smallest variance subject to the unbiasedness condition

Result: the BLUE Estimator for Our DGP

Page 8: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-8

Review: BLUE Estimators (cont.)

• Ordinary Least Squares (OLS) is BLUE for our Gauss–Markov DGP.

• This result is called the “Gauss–Markov Theorem.”

Page 9: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-9

Review: BLUE Estimators (cont.)

• OLS is a very good strategy for the Gauss–Markov DGP.

• OLS is unbiased: our guesses are right on average.

• OLS is efficient: it has a small variance (or at least the smallest possible variance for unbiased linear estimators).

Page 10: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-10

Review: BLUE Estimators (cont.)

• Our guesses will tend to be close to right (or at least as close to right as we can get).

• Warning: the minimum variance could still be pretty large!

Page 11: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-11

Gauss–Markov with an Intercept

Yi

0

1X

i

i (i 1...n)

E(i) 0

Var(i) 2

Cov(i,

j) 0, i j

X 's fixed across samples.

All we have done is add a 0.

Page 12: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-12

BLUE Estimator of 1

12

1

( )( )ˆ ( )

i in

jj

X X Y Y

X X

• This estimator is OLS for the DGP with an intercept. It is the Best (minimum variance) Linear Unbiased Estimator for the Gauss–Markov DGP with an intercept.

Page 13: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-13

Example: The Phillips Curve

• The US data from 1958–1969 suggest a trade-off between inflation and unemployment.

0

1

ˆ 0.06

ˆ 0.55

Unemploymentt 0.06 - 0.55·Inflationt

Page 14: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-14

Example: The Phillips Curve (cont.)

• How do we interpret these numbers?

• If Inflation were 0, our best guess of Unemployment would be 0.06 percentage points.

• A one percentage point increase of Inflation decreases our predicted Unemployment level by 0.55 percentage points.

Unemploymentt 0.06 - 0.55·Inflationt

Page 15: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-15

Figure 4.2 U.S. Unemployment and Inflation, 1958–1969

Page 16: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-16

TABLE 4.1 The Phillips Curve

Page 17: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-17

Example: Phillips Curve

• A straight line seems not quite to capture the data.

• A curve would fit better.

• Can we use linear estimators to fit non-linear relationships?

Page 18: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-18

Logarithms in Econometrics (Chapter 4.5)

12

1

( )( )ˆ

( )

i in

jj

X X Y Y

X X

• How can we use a linear estimator to fit a non-linear relationship?

• A trick: do the “nonlinear” part BEFORE plugging into OLS.

Page 19: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-19

Logarithms in Econometrics (cont.)

12

1

( )( )ˆ ( )

i in

jj

X X Y Y

X X

• A trick: we can do something “nonlinear” BEFORE plugging into OLS.

• OLS doesn’t know (or care) where the Y ’s and X ’s came from.

• Instead of plugging X in directly, we can do stuff to it first.

Page 20: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-20

Logarithms in Econometrics (cont.)

• We can do “stuff” to X and Y before we plug them into our OLS formula.

• If we do something nonlinear to Y and/or X, then our linear regression of new-nonlinear-Y on new-nonlinear-X will plot a nonlinear relationship between Y and X.

• What can we do to X and Y to fit a curve?

Page 21: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-21

Logarithms in Econometrics (cont.)

• LOGARITHMS!!!!!!!!!!!!

• Regress

• How do we interpret our estimated coefficients?

• Our fitted value for 1 no longer tells us the effect of a 1-unit change in X on Y. It tells us the effect of a 1-unit change in log(X) on log(Y).

log(Y ) 0 1·log(X)

Page 22: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-22

Logarithms in Econometrics (cont.)

• Regress

• Our fitted value for 1 no longer tells us the effect of a 1-unit change in X on Y. It tells us the effect of a 1-unit change in log(X) on log(Y).

• Unit changes in log-X translate into PERCENTAGE changes in X.

log(Y ) 0 1·log(X)

Page 23: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-23

• Regress

• Unit changes in log(X) translate into PERCENTAGE changes in X.

• Our estimate tells us the percentage change in Y we predict from a 1-percent change in X.

1

Logarithms in Econometrics (cont.)

log(Y ) 0 1·log(X)

Page 24: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-24

• The percentage change in Y from a 1-percent change in X has a special name in economics—elasticity!

• Taking logs of both Y and X lets us estimate elasticities!

Logarithms in Econometrics (cont.)

log(Y ) 0 1·log(X)

Page 25: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-25

Example: The Phillips Curve

• If we log both Unemployment and Inflation, then we can predict the percentage change in unemployment resulting from a one percent change in inflation.

• Percentage changes are nonlinear. Changing inflation from 0.01 to 0.02 is a 100% increase. Changing inflation from 0.02 to 0.03 is only a 50% increase.

Page 26: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-26

Figure 4.6 A Logarithmic Phillips Curve

Page 27: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-27

TABLE 4.2 A Logarithmic Phillips Curve

Page 28: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-28

Example: The Phillips Curve

log(Unemployment) =

-4.20 – 0.30·log(Inflation)

• A 1-unit increase in log(Inflation) leads to a -0.30 unit change in log(Unemployment). Huh?

• A 1-percent increase in Inflation leads to a -0.30% change in Unemployment. Much more natural interpretation.

Page 29: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-29

Example: The Phillips Curve (cont.)

• Caution: do not confuse percentage points changes and percentage changes.

• If inflation is 15%, a 10 percentage point change moves inflation to 25%

• If inflation is 15%, a 10 percent change moves inflation to 16.5%

Page 30: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-30

0 1i i iY X

Residuals (Chapter 5.1)

• So far, we have learned how to predict lines with an intercept and a single explanator.

• We have focused on making our “best guess” estimates of the intercept and slope of the line.

Page 31: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-31

Residuals (cont.)

• Once we have run our regression, we can make our best guess about what the Y value would be for an observation with some specified value of X.

• Our best guesses about the Y values are called “predictions.”

• If we have an unbiased estimator, we expect our predictions to be right on average.

Page 32: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-32

Residuals (cont.)

• Our best guesses about the Y values are called “predictions.”

• If we have an unbiased estimator, we expect our predictions to be right on average.

• For any given observation, our predictions will be WRONG.

Page 33: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-33

Residuals (cont.)

• Our “best guess” of Yi (conditioned on knowing Xi) is:

• On average, we expect a one-unit increase in X to increase Y by units.

0 1ˆ ˆ

i iY X

1

Page 34: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-34

Residuals (cont.)

• Our “best guess” of Yi (conditioned on knowing Xi) is:

• An individual observation, though, is subject to various factors that are outside our model, such as omitted variables or simple chance.

0 1ˆ ˆ

i iY X

Page 35: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-35

0 1ˆ ˆ

i iY X

ˆi iY Y

Residuals (cont.)

• Our “best guess” of Yi (conditioned on knowing Xi) is:

• Because of these other factors not in our model,

Page 36: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-36

Residuals (cont.)

• That our predictions are “wrong” does NOT mean that our model is wrong.

• After all, our model always said there was a stochastic error term in addition to the deterministic component.

• So far, we have only estimated the deterministic part of our model (the ’s).

Page 37: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-37

Residuals (cont.)

• In our DGP, the non-deterministic component of our model is called “the error term.” We often denote it with an .

• In our data, the non-predicted component of an actual Y observation is called “the residual.” We often denote it with an e.

0 1i i iY X

0 1ˆ ˆ ˆ

i i i i iY X e Y e

Page 38: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-38

Figure 5.1 A Fitted Line and Its Residuals

Page 39: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-39

0 1

0 1ˆ ˆ

i i i

i i i

Y X

Y X e

Residuals

• Residuals are, in part, an accounting gimmick. They make our model “add up.”

• Residuals are ALSO a window into the unseen error term of the model.

• If our estimated ’s are “good,” then our estimated residuals mirror the ’s.

Page 40: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-40

Residuals (cont.)

• We can use the residuals to tell us something about the error terms.

• So what? The error terms are random and vary from sample to sample. We already KNOW they average to 0.

• What we DON’T know is the VARIANCE of the error terms, 2.

Page 41: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-41

2

2

1

1( ) ( )

n

i ii

Varn

2 2

1

1( )

n

ii

s e en

Residuals (cont.)

• By convention, instead of using to denote our estimate of the 2, we use s2

• The definition of a variance is

• However, simply plugging ei in place of i would yield a BIASED estimator of s2

Page 42: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-42

Residuals (cont.)

• Dividing by n-k-1 corrects the bias.

• k is the number of X variables in our regression. So far, we have learned only how to deal with the case k = 1

2 2 2

1 1

1 1( )

2 2

n n

i ii i

s e e en n

Page 43: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-43

Residuals (cont.)

2 2

1

20 1

1

1

2

1 ˆ ˆ( )2

n

ii

n

i ii

s en

Y Xn

Page 44: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-44

Residuals (cont.)

• Once we have estimated s2, we can use it to estimate the variance of our estimator.

• Then we can use that estimated variance to compute Confidence Intervals, which are a guide to the precision of our estimates.

Page 45: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-45

Estimating an Estimator’s Variance(Chapter 5.2)

• We chose OLS as our Best Linear Unbiased Estimator because it is unbiased and efficient (for the Gauss–Markov DGP):

– Unbiased: the estimate is “right” in expectation

– Efficient: the estimator has the lowest possible variance

• Our variance is the lowest possible, but what is that variance?

Page 46: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-46

Estimating an Estimator’s Variance (cont.)

• We can calculate the variance of a linear estimator using the algebra of variances and our Gauss–Markov assumptions.

2

2

2 2 2

( ) ( ) 2 Covariance Terms

( ) 0 ( )

( )

0 ( ) 0

i i i i

i i i i

i i i

i i i

Var wY Var wY

Var wY w Var Y

w Var X

w Var w

Page 47: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-47

Example: The Phillips Curve

log(Unemployment) = 0 + 1·log(Inflation)

• A 1-percent increase in Inflation leads on average to a 1% change in Unemployment.

• We estimate our best guess of 1.

Page 48: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-48

TABLE 4.2 A Logarithmic Phillips Curve

Page 49: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-49

Example: The Phillips Curve

• Our “best guess” for the value of each coefficient is:

• With the fitted coefficients, we can– Predict Y values

– Make our best guess of the sign of each – Make our best guess of the magnitude of each

0

1

ˆ 4.20

ˆ 0.30

Page 50: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-50

Example: The Phillips Curve (cont.)

• Our “best guess” for the value of each coefficient is:

• But how good are these guesses?

0

1

ˆ 4.20

ˆ 0.30

Page 51: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-51

Example: The Phillips Curve (cont.)

• We cannot know exactly how good our estimates are, because we do not know the true parameter values (or else we wouldn’t need to estimate them!)

• We know our estimates are right “on average.”

• To gauge how far off our individual estimator could be, we need to estimate the variance of the linear estimator.

Page 52: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-52

Estimating an Estimator’s Variance(Chapter 5.2)

1

2 2

2

2ˆ 2

2

22

22

2

2

( )

ˆ .

ˆ( ) )

i i i

ii i i

i

ii i

i

i

j

j

Var wY w

xw x X X

x

xVar w Y

x

x

x

x

1

1

Remember, for , , where

(for

Page 53: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-53

Estimating an Estimator’s Variance (cont.)

• We can calculate xi2 directly.

• The other component of the variance of the estimator is 2, the variance of the error term.

Page 54: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-54

Estimating an Estimator’s Variance (cont.)

• We do not observe the error terms directly, so we cannot calculate their variance directly

• However, we can proxy for 2 using s2

Page 55: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-55

Estimating an Estimator’s Variance (cont.)

2

1 2

2

1 2

ˆ( )

ˆ( )

i

i

Varx

sVar

x

Estimated

Page 56: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-56

Estimating an Estimator’s Variance (cont.)

• In practice, we typically use not the variance of the estimator, but the standard deviation of the estimator.

• By convention, the standard deviation of an estimator is referred to as the “standard error.”

Page 57: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-57

Estimating an Estimator’s Variance (cont.)

2

1 2

2

1 2

ˆ. .( )

ˆ. .( )

i

i

s ex

ss e

x

Estimated

Page 58: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-58

Checking Understanding

2

2

ˆ

. . .( 1)

i

i

ee s e

n k x

1The formula for the estimated standard error of is

Question: What properties allow for a

low estimated standard error (i.e. a more precise

distribution of our estimator)?

Page 59: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-59

Checking Understanding (cont.)

2

2. . .

( 1)

i

i

ee s e

n k x

1. If the residuals have a lower variance,

e.s.e.’s will be lower. Efficient estimators have the lowest possible ei

2

2. The higher our sample size n, the lower the e.s.e. Bigger samples are better.

Page 60: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-60

Checking Understanding (cont.)

2

2. . .

( 1)

i

i

ee s e

n k x

3. The fewer parameters we’re estimating,

all else equal, the lower our estimated standard errors.

4. The more broadly our X variable is dispersed, the lower our standard errors. We need variation in X to get a tight estimate of its coefficient.

Page 61: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-61

Estimating an Estimator’s Variance

• The estimator for the constant has different weights, so the formula is a little different.

2 2

0 2

2 2

0 2

ˆ. .( )

ˆ. . .( )

i

j

i

j

Xs e

n x

s Xe s e

n x

Page 62: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-62

Example: The Phillips Curve

• What is the variance of our estimators for the log-log Phillips Curve?

• We need to know 3 things:– SSR = 0.292.

– xi2 = 4.032

– n = 12

• The SSR and n are often reported in regression output. To perform this computation by hand, you would probably need to generate the xi

2 separately.

Page 63: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-63

Example: The Phillips Curve (cont.)

2

1 2 22ˆ. . .( )

0.29210 0.085

4.032

i i

SSRs ne s ex x

Page 64: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-64

TABLE 4.2 A Logarithmic Phillips Curve

Page 65: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-65

Estimating an Estimator’s Variance

• Our estimate for the coefficient on log(inflation) is -0.298.

• The standard error of the estimator is 0.085 (or, using the more precise output of the computer, 0.089).

• We want to use the distribution of the estimator to gauge the precision of our estimate. That is, how close could we reasonably believe we’ve gotten to the true coefficient?

Page 66: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-66

Confidence Intervals (Chapter 5.4)

• So far, we have estimated “best guesses.”

• Looking just at “best guesses” though can be a bit misleading. Our “best guess” could be pretty far from the truth.

• We can get a sense of how far off our “best guess” COULD be, by looking at ALL the “good guesses.”

• A “good guess” is a guess with a reasonable chance of being right.

Page 67: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-67

Confidence Intervals (cont.)

• A “Confidence Interval” is a range of plausible parameter values.

• To interpret Confidence Intervals, we need to think of generating many, many samples from our Data Generating Process.

• For each sample, we observe a C.I.

• Our 95% C.I. will contain the “true” value in 95% of the samples.

Page 68: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-68

Confidence Intervals (cont.)

• In 95% of samples, our 95% C.I. will contain the “true value.”

• We could also calculate 90% or 99% C.I.’s (or for any other “confidence level”)

• If the C.I. is “wide” (i.e. it contains many values), then it is plausible that our “best guess” estimate is far away from the true value.

Page 69: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-69

Confidence Intervals (cont.)

• For the Phillips Curve, our “best guess” estimate of the coefficient on log(inflation) is -0.298. A 1% increase in inflation leads to a predicted -0.298% change in unemployment.

• The 95% C.I. is (-0.496, -0.101).

• We could plausibly guess that the effect was -0.496%, -0.101%, or any number in between.

Page 70: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-70

Confidence Intervals (cont.)

• For the Phillips Curve, our “best guess” estimate of the coefficient on log(inflation) is -0.298. The 95% C.I. is (-0.496, -0.101).

• Was our estimate precise or not? In practice, whether or not we have “good” precision depends on how precise we need to be for our application. If we get the same policy recommendation for all values in our C.I., then we don’t care much which value is “right.”

Page 71: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-71

Confidence Intervals (cont.)

• In this case, we probably DO care whether the true value is closer to -0.496 or to -0.101. Decreasing unemployment by 0.496% for a 1% increase in inflation may be a very attractive trade-off, but decreasing it by only 0.101% may be much less attractive.

• Policy makers would probably desire a much more precise estimate.

Page 72: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-72

Confidence Intervals (cont.)

• For the Phillips Curve, our “best guess” estimate of the coefficient on log(inflation) is -0.298. The 95% C.I. is (-0.496, -0.101).

• Our estimate is right in the middle of the C.I.

• We generate the C.I. by starting with our estimate, and adding a (carefully calculated) “fudge factor.”

Page 73: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-73

Confidence Intervals (cont.)

95% ˆ ˆ. . . . .( )C I multiplier e s e 95%{ }

Page 74: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-74

Confidence Intervals (cont.)

• The starting point for our C.I. is our estimate. Our “best guess” must be a “plausible guess.”

• We then add or subtract a (carefully chosen) multiplier times our estimated standard error.

• The smaller our e.s.e., the tighter our C.I.

• Where does the multiplier come from?

Page 75: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-75

Confidence Intervals (cont.)

• Our C.I. comes from the distribution of the estimator.

• The estimator follows the “t” distribution. The shape of the t distribution depends on n-k-1. When n-k-1 is very large, the t distribution is approximately the same as the normal distribution.

Page 76: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-76

Figure SA.11 The Distribution of the t-Statistic Given the Null Hypothesis is True

Page 77: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-77

Confidence Intervals

% 11

2

ˆ ˆ. . . . .( )n kC I t e s e

Page 78: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-78

95% ˆ ˆ. . 1.96 . . .( )C I e s e

Confidence Intervals (cont.)

• When n-k-1 is large:

Page 79: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-79

95%1 1

ˆ ˆ. . 2.23 . . .( )

0.298 2.23 0.089

( 0.496, 0.100)

C I e s e

Confidence Intervals (cont.)

• For our Phillips Curve example, n-k-1 = 10. The multiplier for the 95% level is 2.23. (As n-k-1 gets larger, our multiplier gets smaller, and our C.I. gets tighter.)

Page 80: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-80

• We can perform functions on our variables BEFORE we plug them into OLS.

• Unit changes in log(X) translate into PERCENTAGE changes in X.

• Our estimate tells us the percentage change in Y we predict from a 1-percent change in X.

1

Review

log(Y ) 0 1·log(X)

Page 81: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-81

Review (cont.)

• Our “best guess” of Yi (conditioned on knowing Xi) is

• An individual observation, though, is subject to various factors that are outside our model, such as omitted variables or simple chance.

0 1ˆ ˆ

i iY X

Page 82: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-82

Review (cont.)

• In our DGP, the non-deterministic component of our model is called “the error term.” We often denote it with an .

• In our data, the non-predicted component of an actual Y observation is called “the residual.” We often denote it with an e.

0 1i i iY X

0 1ˆ ˆ ˆ

i i i i iY X e Y e

Page 83: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-83

Review (cont.)

2 2

1

20 1

1

1

1

1 ˆ ˆ( )1

n

ii

n

i ii

s en k

Y Xn k

Page 84: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-84

Review (cont.)

2

1 2

2

1 2

ˆ( )

ˆ( )

i

i

Varx

sVar

x

Estimated

Page 85: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-85

Review (cont.)

• In practice, we typically use not the variance of the estimator, but the standard deviation of the estimator.

• By convention, the standard deviation of an estimator is referred to as the “standard error.”

Page 86: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-86

Review (cont.)

2

1 2

2

1 2

ˆ. .( )

ˆ. .( )

i

i

i i

s ex

ss e

x

x X X

Estimated

where

Page 87: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-87

Review (cont.)

• A “Confidence Interval” is a range of plausible parameter values.

• To interpret Confidence Intervals, we need to think of generating many, many samples from our Data Generating Process.

• For each sample, we observe a C.I.

• Our 95% C.I. will contain the “true” value in 95% of the samples.

Page 88: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-88

Review (cont.)

• In 95% of samples, our 95% Confidence Interval will contain the “true value.”

• We could also calculate 90% or 99% C.I.’s (or for any other “confidence level”)

• If the C.I. is “wide” (i.e. it contains many values), then it is plausible that our “best guess” estimate is far away from the true value.

Page 89: ch04ch05lectr06

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 6-89

Review (cont.)

% 11

2

ˆ ˆ. . . . .( )n kC I t e s e