last time

198
Last Time • Q-Q plots – Q-Q Envelope to understand variation • Applications of Normal Distribution – Population Modeling – Measurement Error • Law of Averages – Part 1: Square root law for s.d. of average – Part 2: Central Limit Theorem Averages tend to normal distribution

Upload: hamish-fleming

Post on 13-Mar-2016

39 views

Category:

Documents


1 download

DESCRIPTION

Last Time. Q-Q plots Q-Q Envelope to understand variation Applications of Normal Distribution Population Modeling Measurement Error Law of Averages Part 1: Square root law for s.d. of average Part 2: Central Limit Theorem Averages tend to normal distribution. Reading In Textbook. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Last Time

Last Time• Q-Q plots

– Q-Q Envelope to understand variation• Applications of Normal Distribution

– Population Modeling– Measurement Error

• Law of Averages– Part 1: Square root law for s.d. of average– Part 2: Central Limit Theorem

Averages tend to normal distribution

Page 2: Last Time

Reading In TextbookApproximate Reading for Today’s Material:

Pages 61-62, 66-70, 59-61, 335-346

Approximate Reading for Next Class:

Pages 322-326, 337-344, 488-498

Page 3: Last Time

Applications of Normal Dist’n1. Population ModelingOften want to make statements about:

The population mean, μ The population standard deviation, σ

Based on a sample from the populationWhich often follows a Normal distribution

Interesting Question:How accurate are estimates?(will develop methods for this)

Page 4: Last Time

Applications of Normal Dist’n2. Measurement ErrorModel measurement

X = μ + eWhen additional accuracy is required,can make several measurements,and average to enhance accuracy

Interesting question: how accurate?(depends on number of observations, …)

(will study carefully & quantitatively)

Page 5: Last Time

Random SamplingUseful model in both settings 1 & 2:

Set of random variables

Assume: a. Independentb. Same distribution

Say: are a “random sample”

nXX ,,1

nXX ,,1

Page 6: Last Time

Law of AveragesLaw of Averages, Part 1:

Averaging increases accuracy,by factor of n

1

Page 7: Last Time

Law of AveragesRecall Case 1:

CAN SHOW:

Law of Averages, Part 2

So can compute probabilities, etc. using:• NORMDIST• NORMINV

n

NX ,~ˆ

,~,,1 NXX n

Page 8: Last Time

Law of AveragesCase 2: any random sampleCAN SHOW, for n “large”

is “roughly” Consequences:• Prob. Histogram roughly mound shaped• Approx. probs using Normal• Calculate probs, etc. using:

– NORMDIST– NORMINV

nXX ,,1

X ,N

Page 9: Last Time

Law of AveragesCase 2: any random sampleCAN SHOW, for n “large”

is “roughly”

Terminology: “Law of Averages, Part 2” “Central Limit Theorem”

(widely used name)

nXX ,,1

X ,N

Page 10: Last Time

Central Limit TheoremFor any random sample

and for n “large”is “roughly”

nXX ,,1

X ,N

Page 11: Last Time

Central Limit TheoremFor any random sample

and for n “large”is “roughly”

Some nice illustrations

nXX ,,1

X ,N

Page 12: Last Time

Central Limit TheoremFor any random sample

and for n “large”is “roughly”

Some nice illustrations:• Applet by Webster West & Todd Ogden

nXX ,,1

X ,N

Page 13: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll single die

Page 14: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll single die

For 10 plays

Page 15: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll single die

For 10 plays,histogram

Page 16: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll single die

For 20 plays,histogram

Page 17: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll single die

For 50 plays,histogram

Page 18: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll single die

For 100 plays,histogram

Page 19: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll single die

For 1000 plays,histogram

Page 20: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll single die

For 10,000 plays,histogram

Page 21: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll single die

For 100,000 plays,histogram

Stabilizes atUniform

Page 22: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll 5 dice

For 1 play,histogram

Page 23: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll 5 dice

For 10 plays,histogram

Page 24: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll 5 dice

For 100 plays,histogram

Page 25: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll 5 dice

For 1000 plays,histogram

Page 26: Last Time

Central Limit TheoremIllustration: West – Ogden Applethttp://www.amstat.org/publications/jse/v6n3/applets/CLT.html

Roll 5 dice

For 10,000 plays,histogram

Looks moundshaped

Page 27: Last Time

Central Limit TheoremFor any random sample

and for n “large”is “roughly”

Some nice illustrations:• Applet by Webster West & Todd Ogden• Applet from Rice Univ.

nXX ,,1

X ,N

Page 28: Last Time

Central Limit TheoremIllustration: Rice Univ. Applethttp://www.ruf.rice.edu/~lane/stat_sim/sampling_dist/index.html

Starting Distribut’n

Page 29: Last Time

Central Limit TheoremIllustration: Rice Univ. Applethttp://www.ruf.rice.edu/~lane/stat_sim/sampling_dist/index.html

Starting Distribut’n user input

Page 30: Last Time

Central Limit TheoremIllustration: Rice Univ. Applethttp://www.ruf.rice.edu/~lane/stat_sim/sampling_dist/index.html

Starting Distribut’n user input(very non-Normal)

Page 31: Last Time

Central Limit TheoremIllustration: Rice Univ. Applethttp://www.ruf.rice.edu/~lane/stat_sim/sampling_dist/index.html

Starting Distribut’n user input(very non-Normal)

Dist’n of average of n = 2

Page 32: Last Time

Central Limit TheoremIllustration: Rice Univ. Applethttp://www.ruf.rice.edu/~lane/stat_sim/sampling_dist/index.html

Starting Distribut’n user input(very non-Normal)

Dist’n of average of n = 2(slightly more mound shaped?)

Page 33: Last Time

Central Limit TheoremIllustration: Rice Univ. Applethttp://www.ruf.rice.edu/~lane/stat_sim/sampling_dist/index.html

Starting Distribut’n user input(very non-Normal)

Dist’n of average of n = 5(little more mound shaped?)

Page 34: Last Time

Central Limit TheoremIllustration: Rice Univ. Applethttp://www.ruf.rice.edu/~lane/stat_sim/sampling_dist/index.html

Starting Distribut’n user input(very non-Normal)

Dist’n of average of n = 10(much more mound shaped?)

Page 35: Last Time

Central Limit TheoremIllustration: Rice Univ. Applethttp://www.ruf.rice.edu/~lane/stat_sim/sampling_dist/index.html

Starting Distribut’n user input(very non-Normal)

Dist’n of average of n = 25(seems very mound shaped?)

Page 36: Last Time

Central Limit TheoremFor any random sample

and for n “large”is “roughly”

Some nice illustrations:• Applet by Webster West & Todd Ogden• Applet from Rice Univ.• Stats Portal Applet

nXX ,,1

X ,N

Page 37: Last Time

Central Limit TheoremIllustration: StatsPortal. Applet

http://courses.bfwpub.com/ips6e

Density for averageof n = 1, from Exponential dist’n

Page 38: Last Time

Central Limit TheoremIllustration: StatsPortal. Applet

http://courses.bfwpub.com/ips6e

Density for averageof n = 1, from Exponential dist’n

Best fit Normaldensity

Page 39: Last Time

Central Limit TheoremIllustration: StatsPortal. Applet

http://courses.bfwpub.com/ips6e

Density for averageof n = 2, from Exponential dist’n

Best fit Normaldensity

Page 40: Last Time

Central Limit TheoremIllustration: StatsPortal. Applet

http://courses.bfwpub.com/ips6e

Density for averageof n = 4, from Exponential dist’n

Best fit Normaldensity

Page 41: Last Time

Central Limit TheoremIllustration: StatsPortal. Applet

http://courses.bfwpub.com/ips6e

Density for averageof n = 10, from Exponential dist’n

Best fit Normaldensity

Page 42: Last Time

Central Limit TheoremIllustration: StatsPortal. Applet

http://courses.bfwpub.com/ips6e

Density for averageof n = 30, from Exponential dist’n

Best fit Normaldensity

Page 43: Last Time

Central Limit TheoremIllustration: StatsPortal. Applet

http://courses.bfwpub.com/ips6e

Density for averageof n = 100, from Exponential dist’n

Best fit Normaldensity

Page 44: Last Time

Central Limit TheoremIllustration: StatsPortal. Applet

http://courses.bfwpub.com/ips6e

Very strongConvergence

For n = 100

Page 45: Last Time

Central Limit TheoremIllustration: StatsPortal. Applet

http://courses.bfwpub.com/ips6e

Looks “pretty good”

For n = 30

Page 46: Last Time

Central Limit TheoremFor any random sample

and for n “large”is “roughly”

How large n is needed?

nXX ,,1

X ,N

Page 47: Last Time

Central Limit TheoremHow large n is needed?

Page 48: Last Time

Central Limit TheoremHow large n is needed?• Depends completely on setup

Page 49: Last Time

Central Limit TheoremHow large n is needed?• Depends completely on setup• Indiv. obs. Close to normal small OK

Page 50: Last Time

Central Limit TheoremHow large n is needed?• Depends completely on setup• Indiv. obs. Close to normal small OK• But can be large in extreme cases

Page 51: Last Time

Central Limit TheoremHow large n is needed?• Depends completely on setup• Indiv. obs. Close to normal small OK• But can be large in extreme cases• Many people “often feel good”,

when n ≥ 30

Page 52: Last Time

Central Limit TheoremHow large n is needed?• Depends completely on setup• Indiv. obs. Close to normal small OK• But can be large in extreme cases• Many people “often feel good”,

when n ≥ 30

Review earlier examples

Page 53: Last Time

Central Limit TheoremIllustration: Rice Univ. Applethttp://www.ruf.rice.edu/~lane/stat_sim/sampling_dist/index.html

Starting Distribut’n user input(very non-Normal)

Dist’n of average of n = 25(seems very mound shaped?)

Page 54: Last Time

Central Limit TheoremIllustration: StatsPortal. Applet

http://courses.bfwpub.com/ips6e

Looks “pretty good”

For n = 30

Page 55: Last Time

Extreme Case of CLTI.e. of: averages ~ Normal,

when individuals are not

Page 56: Last Time

Extreme Case of CLTI.e. of: averages ~ Normal,

when individuals are not

Page 57: Last Time

Extreme Case of CLTI.e. of: averages ~ Normal,

when individuals are not

ppwppw

X i 1..0..1

~

Page 58: Last Time

Extreme Case of CLTI.e. of: averages ~ Normal,

when individuals are not

• I. e. toss a coin: 1 if Head, 0 if Tail

ppwppw

X i 1..0..1

~

Page 59: Last Time

Extreme Case of CLTI.e. of: averages ~ Normal,

when individuals are not

• I. e. toss a coin: 1 if Head, 0 if Tail• Called “Bernoulli Distribution”

ppwppw

X i 1..0..1

~

Page 60: Last Time

Extreme Case of CLTI.e. of: averages ~ Normal,

when individuals are not

• I. e. toss a coin: 1 if Head, 0 if Tail• Called “Bernoulli Distribution”• Individuals far from normal• Consider sample:

ppwppw

X i 1..0..1

~

nXX ,,1

Page 61: Last Time

Extreme Case of CLTBernoulli sample: nXX ,,1

Page 62: Last Time

Extreme Case of CLTBernoulli sample:

(Recall: independent,with same distribution)

nXX ,,1

Page 63: Last Time

Extreme Case of CLTBernoulli sample:

Note: Xi ~ Binomial(1,p)

nXX ,,1

Page 64: Last Time

Extreme Case of CLTBernoulli sample:

Note: Xi ~ Binomial(1,p)

(Count # H’s in 1 trial)

nXX ,,1

Page 65: Last Time

Extreme Case of CLTBernoulli sample:

Note: Xi ~ Binomial(1,p)

So:EXi = p

nXX ,,1

Page 66: Last Time

Extreme Case of CLTBernoulli sample:

Note: Xi ~ Binomial(1,p)

So:EXi = p

Recall np, with p = 1

nXX ,,1

Page 67: Last Time

Extreme Case of CLTBernoulli sample:

Note: Xi ~ Binomial(1,p)

So:EXi = p

var(Xi) = p(1-p)

nXX ,,1

Page 68: Last Time

Extreme Case of CLTBernoulli sample:

Note: Xi ~ Binomial(1,p)

So:EXi = p

var(Xi) = p(1-p)

Recall np(1-p), with p = 1

nXX ,,1

Page 69: Last Time

Extreme Case of CLTBernoulli sample:

Note: Xi ~ Binomial(1,p)

So:EXi = p

var(Xi) = p(1-p)

sd(Xi) = sqrt(p(1-p))

nXX ,,1

Page 70: Last Time

Extreme Case of CLTBernoulli sample:

EXi = p

sd(Xi) = sqrt(p(1-p))

nXX ,,1

Page 71: Last Time

Extreme Case of CLTBernoulli sample:

EXi = p

sd(Xi) = sqrt(p(1-p))

So Law of Averages

nXX ,,1

Page 72: Last Time

Extreme Case of CLTBernoulli sample:

EXi = p

sd(Xi) = sqrt(p(1-p))

So Law of Averages

(a.k.a. Central Limit Theorem)

nXX ,,1

Page 73: Last Time

Extreme Case of CLTBernoulli sample:

EXi = p

sd(Xi) = sqrt(p(1-p))

So Law of Averages gives:

roughly

nXX ,,1

X npppN 1,

Page 74: Last Time

Extreme Case of CLTLaw of Averages:

roughlyX npppN 1,

Page 75: Last Time

Extreme Case of CLTLaw of Averages:

roughly

Looks familiar?

X npppN 1,

Page 76: Last Time

Extreme Case of CLTLaw of Averages:

roughly

Looks familiar? Recall: For X ~ Binomial(n,p) (counts)

X npppN 1,

Page 77: Last Time

Extreme Case of CLTLaw of Averages:

roughly

Looks familiar? Recall: For X ~ Binomial(n,p) (counts) Sample proportion:

X npppN 1,

nXp ˆ

Page 78: Last Time

Extreme Case of CLTLaw of Averages:

roughly

Looks familiar? Recall: For X ~ Binomial(n,p) (counts) Sample proportion: Has: &

X npppN 1,

nXp ˆ

pEpE nX ˆ

npppsd 1ˆ

Page 79: Last Time

Extreme Case of CLTLaw of Averages:

roughly

Finish Connection:

X npppN 1,

Page 80: Last Time

Extreme Case of CLTLaw of Averages:

roughly

Finish Connection: = # of 1’s among i.e. counts up H’s in n trials So ~ Binomial(n,p) And thus:

X npppN 1,

nXX ,,1

n

i iX1

n

i iX1pXX n

i in ˆ1

1

Page 81: Last Time

Extreme Case of CLTLaw of Averages:

roughly

Finish Connection: = # of 1’s among i.e. counts up H’s in n trials So ~ Binomial(n,p) And thus:

X npppN 1,

nXX ,,1

n

i iX1

n

i iX1pXX n

i in ˆ1

1

Page 82: Last Time

Extreme Case of CLTConsequences:

roughlyp npppN 1,

Page 83: Last Time

Extreme Case of CLTConsequences:

roughly

roughly

p npppN 1,

X pnpnpN 1,

Page 84: Last Time

Extreme Case of CLTConsequences:

roughly

roughly

(using and multiply through by n)

p npppN 1,

X pnpnpN 1,

nXp ˆ

Page 85: Last Time

Extreme Case of CLTConsequences:

roughly

roughly

Terminology: Called The Normal Approximation to the Binomial

p npppN 1,

X pnpnpN 1,

Page 86: Last Time

Extreme Case of CLTConsequences:

roughly

roughly

Terminology: Called The Normal Approximation to the Binomial

p npppN 1,

X pnpnpN 1,

Page 87: Last Time

Extreme Case of CLTConsequences:

roughly

roughly

Terminology: Called The Normal Approximation to the Binomial

(and the sample proportion case)

p npppN 1,

X pnpnpN 1,

Page 88: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

Page 89: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): Control n

Page 90: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): Control n Control p

Page 91: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): Control n Control pSee Prob. Histo.

Page 92: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): Control n Control pSee Prob. Histo.

Compare to fit (by mean & sd) Normal dist’n

Page 93: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): n = 20 p = 0.5

Expect: 20*0.5 = 10(most likely)

Page 94: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): n = 20 p = 0.5

Reasonable Normal fit

Page 95: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): n = 20 p = 0.2

Expect: 20*0.2 = 4(most likely)

Page 96: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): n = 20 p = 0.2

Reasonable fit?Not so good at edge?

Page 97: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): n = 20 p = 0.1

Expect: 20*0.1 = 2(most likely)

Page 98: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): n = 20 p = 0.1

Poor Normal fit,Especially at edge?

Page 99: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): n = 20 p = 0.05

Expect: 20*0.05 = 1(most likely)

Page 100: Last Time

Normal Approx. to BinomialExample: from StatsPortal

http://courses.bfwpub.com/ips6e.php

For Bi(n,p): n = 20 p = 0.05

Normal approxIs very poor

Page 101: Last Time

Normal Approx. to BinomialSimilar behavior for p 1:For Bi(n,p): n = 20 p = 0.5

Page 102: Last Time

Normal Approx. to BinomialSimilar behavior for p 1:For Bi(n,p): n = 20 p = 0.8

(1-p) = 0.2

Page 103: Last Time

Normal Approx. to BinomialSimilar behavior for p 1:For Bi(n,p): n = 20 p = 0.9

(1-p) = 0.1

Page 104: Last Time

Normal Approx. to BinomialSimilar behavior for p 1:For Bi(n,p): n = 20 p = 0.95

(1-p) = 0.05

Mirror imageof above

Page 105: Last Time

Normal Approx. to BinomialNow fix p, and let n vary:For Bi(n,p): n = 1 p = 0.3

Page 106: Last Time

Normal Approx. to BinomialNow fix p, and let n vary:For Bi(n,p): n = 3 p = 0.3

Page 107: Last Time

Normal Approx. to BinomialNow fix p, and let n vary:For Bi(n,p): n = 10 p = 0.3

Page 108: Last Time

Normal Approx. to BinomialNow fix p, and let n vary:For Bi(n,p): n = 30 p = 0.3

Page 109: Last Time

Normal Approx. to BinomialNow fix p, and let n vary:For Bi(n,p): n = 100 p = 0.3

Normal Approx.Improves

Page 110: Last Time

Normal Approx. to BinomialHW: C20For X ~ Bi(n,0.25), find:a. P{X < (n/4)+(sqrt(n)/4)}, by BINOMDISTb. P{X ≤ (n/4)+(sqrt(n)/4)}, by BINOMDISTc. P{X ≤ (n/4)+(sqrt(n)/4)}, using the Normal

Approxim’n to the Binomial (NORMDIST),For n = 16, 64, 256, 1024, 4098.

Page 111: Last Time

Normal Approx. to BinomialHW: C20Numerical answers:

n 16 64 256 1024 4096

(a) 0.630 0.674 0.696 0.707 0.713

(b) 0.810 0.768 0.744 0.731 0.725

(c) 0.718 0.718 0.718 0.718 0.718

Page 112: Last Time

Normal Approx. to BinomialHW: C20Numerical answers:

Notes:• Values stabilize over n

(since cutoff = mean + Z sd)

n 16 64 256 1024 4096

(a) 0.630 0.674 0.696 0.707 0.713

(b) 0.810 0.768 0.744 0.731 0.725

(c) 0.718 0.718 0.718 0.718 0.718

Page 113: Last Time

Normal Approx. to BinomialHW: C20Numerical answers:

Notes:• Values stabilize over n• Normal approx. between others• Everything close for larger n

n 16 64 256 1024 4096

(a) 0.630 0.674 0.696 0.707 0.713

(b) 0.810 0.768 0.744 0.731 0.725

(c) 0.718 0.718 0.718 0.718 0.718

Page 114: Last Time

Normal Approx. to BinomialHW: C20Numerical answers:

Notes:• Values stabilize over n• Normal approx. between others

n 16 64 256 1024 4096

(a) 0.630 0.674 0.696 0.707 0.713

(b) 0.810 0.768 0.744 0.731 0.725

(c) 0.718 0.718 0.718 0.718 0.718

Page 115: Last Time

Normal Approx. to BinomialHW: C20Numerical answers:

Notes:• Values stabilize over n• Normal approx. between others

n 16 64 256 1024 4096

(a) 0.630 0.674 0.696 0.707 0.713

(b) 0.810 0.768 0.744 0.731 0.725

(c) 0.718 0.718 0.718 0.718 0.718

Page 116: Last Time

Normal Approx. to BinomialHow large n?

Page 117: Last Time

Normal Approx. to BinomialHow large n?• Bigger is better

Page 118: Last Time

Normal Approx. to BinomialHow large n?• Bigger is better• Could use “n ≥ 30” rule from above

Law of Averages

Page 119: Last Time

Normal Approx. to BinomialHow large n?• Bigger is better• Could use “n ≥ 30” rule from above

Law of Averages• But clearly depends on p

Page 120: Last Time

Normal Approx. to BinomialHow large n?• Bigger is better• Could use “n ≥ 30” rule from above

Law of Averages• But clearly depends on p

– Worse for p ≈ 0

Page 121: Last Time

Normal Approx. to BinomialHow large n?• Bigger is better• Could use “n ≥ 30” rule from above

Law of Averages• But clearly depends on p

– Worse for p ≈ 0– And for p ≈ 1

Page 122: Last Time

Normal Approx. to BinomialHow large n?• Bigger is better• Could use “n ≥ 30” rule from above

Law of Averages• But clearly depends on p

– Worse for p ≈ 0– And for p ≈ 1– i.e. (1 – p) ≈ 0

Page 123: Last Time

Normal Approx. to BinomialHow large n?• Bigger is better• Could use “n ≥ 30” rule from above

Law of Averages• But clearly depends on p• Textbook Rule:

OK when {np ≥ 10 & n(1-p) ≥ 10}

Page 124: Last Time

Normal Approx. to BinomialHW: 5.18 (a. population too small, b. np =

2 < 10)

C21: Which binomial distributions admit a “good” normal approximation?

a. Bi(30, 0.3)b. Bi(40, 0.4)c. Bi(20,0.5)d. Bi(30,0.7)

(no, yes, yes, no)

Page 125: Last Time

And now for somethingcompletely different….

A statistics professor was describing sampling theory to his class, explaining how a sample can be studied and used to generalize to a population.

One of the students in the back of the room kept shaking his head.

Page 126: Last Time

And now for somethingcompletely different….

"What's the matter?" asked the professor. "I don't believe it," said the student, "why not

study the whole population in the first place?"

The professor continued explaining the ideas of random and representative samples.

The student still shook his head.

Page 127: Last Time

And now for somethingcompletely different….

The professor launched into the mechanics of proportional stratified samples, randomized cluster sampling, the standard error of the mean, and the central limit theorem.

The student remained unconvinced saying, "Too much theory, too risky, I couldn't trust just a few numbers in place of ALL of them."

Page 128: Last Time

And now for somethingcompletely different….

Attempting a more practical example, the professor then explained the scientific rigor and meticulous sample selection of the Nielsen television ratings which are used to determine how multiple millions of advertising dollars are spent.

The student remained unimpressed saying, "You mean that just a sample of a few thousand can tell us exactly what over 250 MILLION people are doing?"

Page 129: Last Time

And now for somethingcompletely different….

Finally, the professor, somewhat disgruntled with the skepticism, replied,

"Well, the next time you go to the campus clinic and they want to do a blood test...tell them that's not good enough ...

tell them to TAKE IT ALL!!"

From: GARY C. RAMSEYER• http://www.ilstu.edu/~gcramsey/Gallery.html

Page 130: Last Time

Central Limit TheoremFurther Consequences of Law of Averages:

Page 131: Last Time

Central Limit TheoremFurther Consequences of Law of Averages:1. N(μ,σ) distribution is a useful model for

populations

Page 132: Last Time

Central Limit TheoremFurther Consequences of Law of Averages:1. N(μ,σ) distribution is a useful model for

populationse.g. SAT scores are averages of scores from many

questions

Page 133: Last Time

Central Limit TheoremFurther Consequences of Law of Averages:1. N(μ,σ) distribution is a useful model for

populationse.g. SAT scores are averages of scores from many

questionse.g. heights are influenced by many small factors,

your height is sum of these.

Page 134: Last Time

Central Limit TheoremFurther Consequences of Law of Averages:1. N(μ,σ) distribution is a useful model for

populationse.g. SAT scores are averages of scores from many

questionse.g. heights are influenced by many small factors,

your height is sum of these.

2. N(μ,σ) distribution useful for modeling measurement error

Page 135: Last Time

Central Limit TheoremFurther Consequences of Law of Averages:1. N(μ,σ) distribution is a useful model for

populationse.g. SAT scores are averages of scores from many

questionse.g. heights are influenced by many small factors,

your height is sum of these.

2. N(μ,σ) distribution useful for modeling measurement error

Sum of many small components

Page 136: Last Time

Central Limit TheoremFurther Consequences of Law of Averages:1. N(μ,σ) distribution is a useful model for

populations2. N(μ,σ) distribution useful for modeling

measurement error

Page 137: Last Time

Central Limit TheoremFurther Consequences of Law of Averages:1. N(μ,σ) distribution is a useful model for

populations2. N(μ,σ) distribution useful for modeling

measurement error

Now have powerful probability tools for

Page 138: Last Time

Central Limit TheoremFurther Consequences of Law of Averages:1. N(μ,σ) distribution is a useful model for

populations2. N(μ,σ) distribution useful for modeling

measurement error

Now have powerful probability tools for:a. Political Polls

Page 139: Last Time

Central Limit TheoremFurther Consequences of Law of Averages:1. N(μ,σ) distribution is a useful model for

populations2. N(μ,σ) distribution useful for modeling

measurement error

Now have powerful probability tools for:a. Political Pollsb. Populations – Measurement Error

Page 140: Last Time

Course Big PictureNow have powerful probability tools for:a. Political Pollsb. Populations – Measurement Error

Page 141: Last Time

Course Big PictureNow have powerful probability tools for:a. Political Pollsb. Populations – Measurement Error

Next deal systematically with unknown p & μ

Page 142: Last Time

Course Big PictureNow have powerful probability tools for:a. Political Pollsb. Populations – Measurement Error

Next deal systematically with unknown p & μ

Subject called “Statistical Inference”

Page 143: Last Time

Statistical InferenceIdea: Develop formal framework for

handling unknowns p & μ

Page 144: Last Time

Statistical InferenceIdea: Develop formal framework for

handling unknowns p & μ

(major work for this is already done)

Page 145: Last Time

Statistical InferenceIdea: Develop formal framework for

handling unknowns p & μ

(major work for this is already done)

(now will just formalize, and refine)

Page 146: Last Time

Statistical InferenceIdea: Develop formal framework for

handling unknowns p & μ

(major work for this is already done)

(now will just formalize, and refine)

(do for simultaneously for major models)

Page 147: Last Time

Statistical InferenceIdea: Develop formal framework for

handling unknowns p & μ

e.g. 1: Political Polls

Page 148: Last Time

Statistical InferenceIdea: Develop formal framework for

handling unknowns p & μ

e.g. 1: Political Polls(estimate p = proportion for A)

Page 149: Last Time

Statistical InferenceIdea: Develop formal framework for

handling unknowns p & μ

e.g. 1: Political Polls(estimate p = proportion for A)

(in population, based on sample)

Page 150: Last Time

Statistical InferenceIdea: Develop formal framework for

handling unknowns p & μ

e.g. 1: Political Polls

e.g. 2a: Population Modeling

Page 151: Last Time

Statistical InferenceIdea: Develop formal framework for

handling unknowns p & μ

e.g. 1: Political Polls

e.g. 2a: Population Modeling(e.g. heights, SAT scores …)

Page 152: Last Time

Statistical InferenceIdea: Develop formal framework for

handling unknowns p & μ

e.g. 1: Political Polls

e.g. 2a: Population Modeling

e.g. 2b: Measurement Error

Page 153: Last Time

Statistical InferenceA parameter is a numerical feature

Page 154: Last Time

Statistical InferenceA parameter is a numerical feature of

population

Page 155: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

Page 156: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

(so far parameters have been indicesof probability distributions)

Page 157: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

(so far parameters have been indicesof probability distributions)

(this is an additional role for that term)

Page 158: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 1, Political Polls• Population is all voters

Page 159: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 1, Political Polls• Population is all voters• Parameter is proportion of population for A

Page 160: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 1, Political Polls• Population is all voters• Parameter is proportion of population for A,

often denoted by p

Page 161: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 1, Political Polls• Population is all voters• Parameter is proportion of population for A,

often denoted by p

(same as p before, just new framework)

Page 162: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 2a, Population Modeling• Parameters are μ & σ

Page 163: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 2a, Population Modeling• Parameters are μ & σ

(population mean & sd)

Page 164: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 2b, Measurement Error• Population is set of all possible

measurements

Page 165: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 2b, Measurement Error• Population is set of all possible

measurements

(from thought experiment viewpoint)

Page 166: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 2b, Measurement Error• Population is set of all possible

measurements• Parameters are

Page 167: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 2b, Measurement Error• Population is set of all possible

measurements• Parameters are:

– μ = true value

Page 168: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sample

E.g. 2b, Measurement Error• Population is set of all possible

measurements• Parameters are:

– μ = true value– σ = s.d. of measurements

Page 169: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sampleAn estimate of a parameter is some function

of data

Page 170: Last Time

Statistical InferenceA parameter is a numerical feature of

population, not sampleAn estimate of a parameter is some function

of data

(hopefully close to parameter)

Page 171: Last Time

Statistical InferenceAn estimate of a parameter is some function

of data

E.g. 1: Political PollsEstimate population proportion, p, by

sample proportion: nXp ˆ

Page 172: Last Time

Statistical InferenceAn estimate of a parameter is some function

of data

E.g. 1: Political PollsEstimate population proportion, p, by

sample proportion:

(same as before)

nXp ˆ

Page 173: Last Time

Statistical InferenceAn estimate of a parameter is some function

of data

E.g. 2a,b: Estimate population:mean μ, by sample mean: X

Page 174: Last Time

Statistical InferenceAn estimate of a parameter is some function

of data

E.g. 2a,b: Estimate population:mean μ, by sample mean:s.d. σ, by sample sd:

Xs

Page 175: Last Time

Statistical InferenceAn estimate of a parameter is some function

of data

E.g. 2a,b: Estimate population:mean μ, by sample mean:s.d. σ, by sample sd:

Parameters

Xs

Page 176: Last Time

Statistical InferenceAn estimate of a parameter is some function

of data

E.g. 2a,b: Estimate population:mean μ, by sample mean:s.d. σ, by sample sd:

Estimates

Xs

Page 177: Last Time

Statistical InferenceHow well does an estimate work?

Page 178: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

Page 179: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

(i.e. on average get right answer)

Page 180: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

E.g. 1: pE ˆ

Page 181: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

E.g. 1: nXEpE ˆ

Page 182: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

E.g. 1: nnp

nXEpE ˆ

Page 183: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

E.g. 1: pEpE nnp

nX ˆ

Page 184: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

E.g. 1:

(conclude sample proportion is unbiased)

pEpE nnp

nX ˆ

Page 185: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

E.g. 1:

(conclude sample proportion is unbiased)(i.e. centered correctly)

pEpE nnp

nX ˆ

Page 186: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

E.g. 2a,b: E

Page 187: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

E.g. 2a,b: XEE

Page 188: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

E.g. 2a,b: XEE ˆ

Page 189: Last Time

Statistical InferenceHow well does an estimate work?

Unbiasedness: Good estimate should be centered at right value

E.g. 2a,b:

(conclude sample mean is unbiased)(i.e. centered correctly)

XEE ˆ

Page 190: Last Time

Statistical InferenceHow well does an estimate work?

Standard Error: for an unbiased estimator, standard error is standard deviation

Page 191: Last Time

Statistical InferenceHow well does an estimate work?

Standard Error: for an unbiased estimator, standard error is standard deviation

E.g. 1: SE of is npppsd 1ˆp

Page 192: Last Time

Statistical InferenceHow well does an estimate work?

Standard Error: for an unbiased estimator, standard error is standard deviation

E.g. 2a,b: SE of is n

Xsdsd ˆ

Page 193: Last Time

Statistical InferenceHow well does an estimate work?

Standard Error: for an unbiased estimator, standard error is standard deviation

Same ideas as above

Page 194: Last Time

Statistical InferenceHow well does an estimate work?

Standard Error: for an unbiased estimator, standard error is standard deviation

Same ideas as above:• Gets better for bigger n

Page 195: Last Time

Statistical InferenceHow well does an estimate work?

Standard Error: for an unbiased estimator, standard error is standard deviation

Same ideas as above:• Gets better for bigger n• By factor of n

Page 196: Last Time

Statistical InferenceHow well does an estimate work?

Standard Error: for an unbiased estimator, standard error is standard deviation

Same ideas as above:• Gets better for bigger n• By factor of• Only new terminology

n

Page 197: Last Time

Statistical InferenceNice graphic on bias and variability:

Figure 3.14

From text

Page 198: Last Time

Statistical InferenceHW: C22: Estimate the standard error of:a. The estimate of the population proportion,

p, when the sample proportion is 0.9, based on a sample of size 100. (0.03)

b. The estimate of the population mean, μ, when the sample standard deviation is s=15, based on a sample of size 25 (3)