ams 572 class noteszhu/ams57208/ch12_handout.doc  · web vieweach of the three groups has received...

27
AMS 572 Class Notes Nov . 20, 2007 (b) Chapter 12 Analysis of Variance (ANOVA) One-way ANOVA (fixed factors) * Goal: compare the means from a (a≥2) different populations. * It is an extension of the pooled variance t-test. * Assumptions: (i) Equal (unknown) population variances (ii) Normal populations (iii) Independent samples : these ’s are not all equal. Assumptions: a population, , i=1,2,…,a. is unknown. Samples: a independent samples. Data:

Upload: others

Post on 06-Nov-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

AMS 572 Class Notes

AMS 572 Class Notes

Nov. 20, 2007 (b)

Chapter 12 Analysis of Variance (ANOVA)

One-way ANOVA (fixed factors)

012

:

a

H

mmm

===

L

* Goal: compare the means from a (a≥2) different populations.

* It is an extension of the pooled variance t-test.

* Assumptions:

(i) Equal (unknown) population variances

2222

12

a

ssss

====

L

(ii) Normal populations

(iii) Independent samples

a

H

: these

i

m

’s are not all equal.

Assumptions: a population,

2

(,)

i

N

ms

, i=1,2,…,a.

2

s

is unknown.

Samples: a independent samples.

Data:

1

1

populationpopulationipopulationa

samplesampleisamplea

¯¯¯

1

11

11

22

12

1

1

ia

ia

ia

ia

n

inan

YY

Y

YY

Y

nnn

Y

YY

ìì

ì

ïï

ï

ïïï

ííí

ïïï

ïïï

î

îî

M

MM

Balanced design:

i

nn

º

Unbalanced design: otherwise

Derivation of the test

(1) PQ, can be derived

(2) * Union-intersection method. Best method for this type of test as in other regression analysis related tests. Please see AMS 570/571 text book, and also the book by G.A.F. Seber: Linear Regression Model, published by John Wiley for details.

(3) LRT (Likelihood Ratio test)

Test Statistic:

(

)

(

)

0

2

11

01,

2

11

()/1

~

()/

i

i

n

a

i

H

ij

aNa

n

a

iji

ij

yya

MSA

FF

MSE

yyNa

×××

==

--

×

==

--

==

--

åå

åå

Total sample size

1

a

i

i

Nn

=

=

å

.

Sample mean:

1

ia

YYY

×××

LL

grand mean

Y

××

2

~(,),1,,1,,

iid

ijii

YNiajn

ms

==

LL

Balanced design:

i

nn

=

2

~(,)

iid

ij

YN

ms

2

1

~(,)

YN

n

s

m

×

,

2

~(,)

i

YN

n

s

m

×

2

2

1

1

2

()

~

/

a

i

i

a

YY

n

c

s

×××

=

-

-

å

2

11

22

(1)

2

()

~

an

iji

ij

anNa

YY

cc

s

×

==

--

-

=

åå

0

2

2

1

1

2

()

~

/

a

i

H

i

a

YY

n

c

s

×××

=

-

-

å

Theorem Let

2

~(,),1,,

iid

i

XNin

ms

=

K

(1)

2

~(,)

XN

n

s

m

(2)

2

2

2

1

1

22

()

(1)

~

n

i

i

n

XX

nS

c

ss

=

-

-

-

=

å

(3)

X

and

2

S

are independent.

Definition

12

1

,

2

/

~

/

kk

Wk

FF

Vk

=

, where

12

22

~,~

kk

WV

cc

and they are independent.

2

2

1

()

()

1

a

ii

i

n

EMSA

a

mm

s

=

-

=+

-

å

,

2

()

EMSE

s

=

.

When

0

H

is true:

0

~1

F

×

, (

i

mm

=

)

a

H

is true:

0

1

F

>

.

Intuitively, we reject

0

H

in favor of

a

H

if

0

FC

>

, where C is determined by the significance level as usual:

0000

(|)(|)

PrejectHHPFCH

a

==³

.

When a=2,

01212

::

a

HH

mmmm

0

12

12

02

12

~

11

H

nn

yy

TT

S

nn

××

+-

-

=

+

Note: If

2

1,

~,~

kk

TtTF

. (One can prove this easily using the definitions of the t- and F-distributions)

If we reject the ANOVA hypothesis, then we should do the pairwise comparisons.

(1)

2

2

a

aa

æö

-

=

ç÷

èø

011202130,(1)/21

:::

aaaa

HHH

mmmmmm

--

===

L

112213,(1)/21

:::

aaaaaaa

HHH

mmmmmm

--

¹¹¹

L

The multiple comparison problem

FWE: (Family wise error rate)

=P(reject at least 1 true null hypothesis)

Tukey’s Studentized Range test (* It is the preferred method to ensure the FWE)

0

:

:

ijij

aijij

H

H

mm

mm

=

¹

At FWE

a

, reject

0

ij

H

if

,,

||

112

*

ij

aNa

ij

ij

yy

q

t

s

nn

a

××

-

-

=>

+

Finally, the name ANOVA came from the partitioning of the variations:

222

111111

()()()

iii

nnn

aaa

ijijii

ijijij

yyyyyy

SSTSSESSA

××××××

======

-=-+-

åååååå

For more details, please refer to the text book.

Attachment: Handout.

Ronald Fisher (1890-1962)

Sir Ronald Aylmer Fisher was a British statistician, evolutionary biologist, and geneticist. He has been described as: “a genius who almost single-handedly created the foundations for modern statistical science” and “the greatest of Darwin's successors”.

Fisher was born in East Finchley in London, to George and Katie Fisher. Although Fisher had very poor eyesight, he was a precocious student, winning the Neeld Medal (a competitive essay in Mathematics) at Harrow School at the age of 16. Because of his poor eyesight, he was tutored in mathematics without the aid of paper and pen, which developed his ability to visualize problems in geometrical terms, as opposed to using algebraic manipulations. He was legendary in being able to produce mathematical results without setting down the intermediate steps. In 1909 he won a scholarship to Gonville and Caius College, Cambridge, and graduated with a degree in mathematics in 1913.

During his work as a statistician at the Rothamsted Agricultural Experiment Station, UK, Fisher pioneered the principles of the design of experiments and elaborated his studies of "analysis of variance". In addition to "analysis of variance", Fisher invented the technique of maximum likelihood and originated the concepts of sufficiency, ancillarity, Fisher's linear discriminator and Fisher information. The contributions Fisher made also included the development of methods suitable for small samples, like those of Gosset, and the discovery of the precise distributions of many sample statistics. Fisher published a number of important texts including Statistical Methods for Research Workers (1925), The design of experiments (1935) and Statistical tables (1947). Fisher's important contributions to both genetics and statistics are emphasized by the remark of L.J. Savage, "I occasionally meet geneticists who ask me whether it is true that the great geneticist R.A. Fisher was also an important statistician" (Annals of Statistics, 1976).

John Tukey (1915-2000)

John Tukey, 85, Statistician; Coined the Word 'Software'

John Wilder Tukey, one of the most influential statisticians of the last 50 years and a wide-ranging thinker credited with inventing the word ''software,'' died on Wednesday in New Brunswick, N.J. He was 85.

The cause was a heart attack after a short illness, said Phyllis Anscombe, his sister-in-law.

Mr. Tukey developed important theories about how to analyze data and compute series of numbers quickly. He spent decades as both a professor at Princeton University and a researcher at AT&T's Bell Laboratories, and his ideas continue to be a part of both doctoral statistics courses and high school math classes. In 1973, President Richard M. Nixon awarded him the National Medal of Science.

But Mr. Tukey frequently ventured outside of the academy as well, working as a consultant to the government and corporations and taking part in social debates.

In the 1950's, he criticized Alfred C. Kinsey's research on sexual behavior. In the 1970's, he was chairman of a research committee that warned that aerosol spray cans damaged the ozone layer. More recently, he recommended that the 1990 Census be adjusted by using statistical formulas in order to count poor urban residents whom he believed it had missed.

''The best thing about being a statistician,'' Mr. Tukey once told a colleague, ''is that you get to play in everyone's backyard.''

An intense man who liked to argue and was fond of helping other researchers, Mr. Tukey was also an amateur linguist who made significant contributions to the language of modern times. In a 1958 article in American Mathematical Monthly, he became the first person to define the programs on which electronic calculators ran, said Fred R. Shapiro, a librarian at Yale Law School who is editing a dictionary of quotations with information on the origin of terms. Three decades before the founding of Microsoft, Mr. Tukey saw that ''software,'' as he called it, was gaining prominence. ''Today,'' he wrote at the time, it is ''at least as important'' as the '' 'hardware' of tubes, transistors, wires, tapes and the like.''

Twelve years earlier, while working at Bell Laboratories, he had coined the term ''bit,'' an abbreviation of ''binary digit'' that described the 1's and 0's that are the basis of computer programs.

Both words caught on, to the chagrin of some computer scientists who saw Mr. Tukey as an outsider. ''Not everyone was happy that he was naming things in their field,'' said Steven M. Schultz, a spokesman for Princeton.

Mr. Tukey had no immediate survivors. His wife of 48 years, Elizabeth Rapp Tukey, an antiques appraiser and preservation activist, died in 1998.

Mr. Tukey was born in 1915 in New Bedford, a fishing town on the southern coast of Massachusetts, and was the only child of Ralph H. Tukey and Adah Tasker Tukey. His mother was the valedictorian of the class of 1898 at Bates College in Lewiston, Me., and her closest competition was her eventual husband, who became the salutatorian. Classmates referred to them as the couple most likely to give birth to a genius, said Marc G. Glass, a Bates spokesman.

The elder Mr. Tukey became a Latin teacher at New Bedford's high school, but, because of a rule barring spouses from teaching at the school, Mrs. Tukey was a private tutor, Mrs. Anscombe said. Mrs. Tukey's main pupil became her son, who attended regular classes only for special subjects like French. ''They were afraid that if he went to school, he'd get lazy,'' said Howard Wainer, a friend and former student of John Tukey's.

In 1936, Mr. Tukey graduated from nearby Brown University with a bachelor's degree in chemistry, and in the next three years earned three graduate degrees, one in chemistry at Brown and two in mathematics at Princeton, where he would spend the rest of his career. At the age of 35, he became a full professor, and in 1965 he became the founding chairman of Princeton's statistics department.

Mr. Tukey worked for the United States government during World War II. Friends said he did not discuss the details of his projects, but Mrs. Anscombe said he helped design the U-2 spy plane.

In later years, much of his important work came in a field that statisticians call robust analysis, which allows researchers to devise credible conclusions even when the data with which they are working are flawed. In 1970, Mr. Tukey published ''Exploratory Data Analysis,'' which gave mathematicians new ways to analyze and present data clearly.

One of those tools, the stem-and-leaf display, continues to be part of many high school curriculums. Using it, students arrange a series of data points in a series of simple rows and columns and can then make judgments about what techniques, like calculating the average or median, would allow them to analyze the information intelligently.

That display was typical of Mr. Tukey's belief that mathematicians, professional or amateur, should often start with their data and then look for a theorem, rather than vice versa, said Mr. Wainer, who is now the principal research scientist at the Educational Testing Service.

''He legitimized that, because he wasn't doing it because he wasn't good at math,'' Mr. Wainer said. ''He was doing it because it was the right thing to do.''

Along with another scientist, James Cooley, Mr. Tukey also developed the Fast Fourier Transform, an algorithm with wide application to the physical sciences. It helps astronomers, for example, determine the spectrum of light coming from a star more quickly than previously possible.

As his career progressed, he also became a hub for other scientists. He was part of a group of Princeton professors that gathered regularly and included Lyman Spitzer Jr., who inspired the Hubble Space Telescope. Mr. Tukey also persuaded a group of the nation's top statisticians to spend a year at Princeton in the early 1970's working together on robust analysis problems, said David C. Hoaglin, a former student of Mr. Tukey.

Mr. Tukey was a consultant to the Educational Testing Service, the Xerox Corporation and Merck & Company. From 1960 to 1980, he helped design the polls that the NBC television network used to predict and analyze elections.

His first brush with publicity came in 1950, when the National Research Council appointed him to a committee to evaluate the Kinsey Report, which shocked many Americans by describing the country's sexual habits as far more diverse than had been thought. From their first meeting, when Mr. Kinsey told Mr. Tukey to stop singing a Gilbert and Sullivan tune aloud while working, the two men clashed, according to ''Alfred C. Kinsey,'' a biography by James H. Jones.

In a series of meetings over two years, Mr. Kinsey vigorously defended his work, which Mr. Tukey believed was seriously flawed, relying on a sample of people who knew each other. Mr. Tukey said a random selection of three people would have been better than a group of 300 chosen by Mr. Kinsey.

By DAVID LEONHARDT, July 28, 2000 © The New York Times Company

Example 1. A deer (definitely not reindeer) hunter prefers to practice with several different rifles before deciding which one to use for hunting. The hunter has chosen five particular rifles to practice with this season. In one test to see which rifles could shoot the farthest and still have sufficient knock-down power, each rifle was fired six times and the distance the bullet traveled recorded. A summary of the sample data is listed below, where the distances are recorded in yards.

Rifle

Mean

Std. Dev.

1

341.7

40.8

2

412.5

23.6

3

365.8

62.2

4

505.0

28.3

5

430.0

38.1

(a) Are these rifles equally good? Test at =0.05.

Answer: This is one-way ANOVA with 5 “samples” (a=5), and 6 observations per sample (

6

i

nn

º=

), and thus the total sample size is N=30. The grand mean is

341.7412.5430.0

411

5

y

+++

==

L

We are testing

0125

:

H

mmm

===

L

versus

a

H

: at least one of these equalities is not true. The test statistic is

0

MSA

F

MSE

=

where

2

22

1

()

6

[(341.7411)(430.0411)]24067.17

14

a

ii

i

nyy

MSA

a

=

-

==-++-=

-

å

K

and

2

222

1

1

(1)

1

[40.838.1]1668.588

5

(1)

a

ii

i

a

i

i

ns

MSEs

n

=

=

-

===++=

-

å

å

K

Therefore

0

24067.17

14.42

1668.588

F

Since

04,25,0.05

14.422.67

FF

»>=

, we reject the ANOVA hypothesis

0

H

and claim that the five rifles are not equally good.

(b) Use Tukey’s procedure with (=0.05 to make pairwise comparisons among the five population means.

Answer: Now we will do the pairwise comparison using Tukey’s method. The Tukey method will reject any pairwise null hypothesis

0

:

ijij

H

mm

=

at FWE=( if

,,

/

ij

aNa

yy

q

sn

a

-

-

>

In our case, a=5, n=6,

1668.58840.85

s

, N-a=25, (=0.05, and

,,5,30,0.05

4.17

aNa

qq

a

-

. Therefore, we would reject

0

ij

H

if

,,

*/69.54

ijaNa

yysqn

a

-

->»

The conclusion is that at the familywise error rate of 0.05, we declare that the following rifle pairs are significantly different: 4/1, 4/2, 4/3, 4/5, 5/1, 2/1.

Example 2. Fifteen subjects were randomly assigned to three treatment groups X, Y and Z (with 5 subjects per treatment). Each of the three groups has received a different method of speed-reading instruction. A reading test is given, and the number of words per minute is recorded for each subject. The following data are collected:

X

Y

Z

700

480

500

850

460

550

820

500

480

640

570

600

920

580

610

Please write a SAS program to answer the following questions.

(a) Are these treatments equally effective? Test at α = 0.05.

(b) If these treatments are not equally good, please use Tukey’s procedure with α = 0.05 to make pairwise comparisons.

Answer: This is one-way ANOVA with 3 samples and 5 observations per sample. The SAS code is as follows:

DATA READING;

INPUT GROUP $ WORDS @@;

DATALINES;

X 700 X 850 X 820 X 640 X 920 Y 480 Y 460 Y 500

Y 570 Y 580 Z 500 Z 550 Z 480 Z 600 Z 610

;

RUN ;

PROC ANOVA DATA=READING ;

TITLE ‘Analysis of Reading Data’ ;

CLASS GROUP;

MODEL WORDS = GROUP;

MEANS GROUP / TUKEY;

RUN;

The SAS output is as follows:

Analysis of Reading Data

The ANOVA Procedure

Class Level Information

Class Levels Values

GROUP 3 X Y Z

Number of observations 15

The ANOVA Procedure

Dependent Variable: WORDS

Sum of

Source DF Squares Mean Square F Value Pr > F

Model 2 215613.3333 107806.6667 16.78 0.0003

Error 12 77080.0000 6423.3333

Corrected Total 14 292693.3333

R-Square Coeff Var Root MSE WORDS Mean

0.736653 12.98256 80.14570 617.3333

Source DF Anova SS Mean Square F Value Pr > F

GROUP 2 215613.3333 107806.6667 16.78 0.0003

The ANOVA Procedure

Tukey's Studentized Range (HSD) Test for WORDS

NOTE: This test controls the Type I experimentwise error rate, but it generally has a higher Type

II error rate than REGWQ.

Alpha 0.05

Error Degrees of Freedom 12

Error Mean Square 6423.333

Critical Value of Studentized Range 3.77278

Minimum Significant Difference 135.22

Means with the same letter are not significantly different.

Tukey Grouping Mean N GROUP

A 786.00 5 X

B 548.00 5 Z

B

B 518.00 5 Y

(a) Conclusion:

(b) The p-value of the ANOVA F-test is 0.0003, less than the significance level α = 0.05. Thus we conclude that the three reading methods are not equally good.

(c) Furthermore, the Tukey’s procedure with α = 0.05 shows that method X is superior to methods Y and Z, while methods Y and Z are not significantly different.

http://www.stat.sfu.ca/~sgchiu/Grace/S330/handouts/1-way/

Example 3. One-Way ANOVA on SAS -- Motor Oil Example. The SAS code:

/****************************************************

A sample SAS program to analyze the Motor Oil data

****************************************************/

title 'Motor Oil analysis';

options nodate noovp linesize=68;

/****************************************************

Virtually all SAS programs consist of a DATA step where

the raw data is read into a SAS file, and procedure (PROC)

step which perform various analyses

****************************************************/

data oil;

/* this will read the data into an internal SAS dataset called 'oil' */

input type $ visc;

/* the $ indicates 'type' is a character variable */

label type='oil type';

label visc='viscosity';

datalines; /* data follows */

CNVNTNL 44

CNVNTNL 49

CNVNTNL 37

CNVNTNL 38

SYNTHET 42

SYNTHET 49

SYNTHET 52

SYNTHET 57

HYBRID 60

HYBRID 58

HYBRID 78

;

proc print data=oil;

/* this will print out the raw data for checking */

title2 'raw data';

proc sort data=oil;

by type;

proc means data=oil maxdec=2 n mean std;

/* get simple summary statistics (sample size, sample mean and SD) with

max of 2 decimal places */

title2 'simple summary statistics';

by type; /* statistics computed for each oil type... */

var visc; /* ... on the variable 'visc' */

proc plot data=oil;

/* request a plot of the raw data */

title2 'plot of the raw data';

plot visc*type;

proc anova data=oil;

title2 'Analysis';

class type;

/* class statement indicates that 'oil type' is a factor */

model visc = type;

/* assumes 'oil type' influences 'viscosity' */

means type / tukey cldiff;

/* multiple comparison by Tukey's method -- get actual C.I.'s */

means type / tukey lines;

/* get pictorial display of comparisons */

proc glm data=oil;

title2 'Proc glm Analysis';

/* same as 'proc anova' except

'glm' allows residual plots but gives more junk output */

class type;

model visc = type;

output out=oilfit p=yhat r=resid;

/* store fitted values and fitted residuals

in dataset called 'oilfit' for later use */

proc univariate data=oilfit plot normal;

var resid;

/* plot qq-plot of fitted residuals */

proc plot;

plot resid*type;

plot resid*yhat;

/* two residual plots to check

independence and constant variance */

run;

Related homework problems:

Chapter 12: 1, 2, 5, 8, 10, 33* (for AMS PhD students only)

Solutions to the above homework problems:

12.1

(a) Sugar :

996

.

3

3

)

865

.

1

(

)

985

.

1

(

)

138

.

2

(

2

2

2

2

=

+

+

=

s

,

So that

999

.

1

996

.

3

=

=

s

with

57

19

3

=

´

d.f. Using the critical value

000

.

2

025

.

0

,

57

»

t

, the 95% CI’s are :

Shelf 1 :

]

694

.

5

,

906

.

3

[

20

999

.

1

)

000

.

2

(

80

.

4

=

´

±

Shelf 2 :

]

744

.

10

,

956

.

8

[

20

999

.

1

)

000

.

2

(

85

.

9

=

´

±

Shelf 3 :

]

994

.

6

,

206

.

5

[

20

999

.

1

)

000

.

2

(

10

.

6

=

´

±

Fiber :

447

.

1

3

)

277

.

1

(

)

162

.

1

(

)

166

.

1

(

2

2

2

2

=

+

+

=

s

,

So that

203

.

1

447

.

1

=

=

s

with

57

19

3

=

´

d.f. Using the critical value

000

.

2

025

.

0

,

57

»

t

, the 95% CI’s are :

Shelf 1 :

]

218

.

2

,

142

.

1

[

20

203

.

1

)

000

.

2

(

68

.

1

=

´

±

Shelf 2 :

]

488

.

1

,

412

.

0

[

20

203

.

1

)

000

.

2

(

95

.

0

=

´

±

Shelf 3 :

]

708

.

2

,

632

.

1

[

20

203

.

1

)

000

.

2

(

17

.

2

=

´

±

Shelf 2 cereals are higher in sugar content than shelves 1 and 3, since the CI for shelf 2 is above those of shelves 1 and 3. Similarly, the shelf 2 fiber content CI is below that of shelf 3. So in general, shelf 2 cereals are higher in sugar and lower in fiber.

(b) Sugar :

80

.

227

996

.

3

.

57

=

´

=

SSE

,

å

-

=

2

2

y

N

y

n

SSA

i

2

2

2

2

)

92

.

6

(

60

]

)

10

.

6

(

)

85

.

9

(

)

80

.

4

[(

20

´

-

+

+

=

03

.

275

=

Then the ANOVA table is below :

Analysis of Variance

Source

SS

d.f.

MS

F

Shelves

275.03

2

137.5

34.41

Error

227.80

57

3.996

Total

502.83

59

Since

15

.

3

05

.

0

,

57

.

2

=

f

F

f

, there do appear to be significant differences among the shelves in terms of sugar content.

Fiber :

47

.

82

447

.

1

.

57

=

´

=

SSE

,

å

-

=

2

2

y

N

y

n

SSA

i

2

2

2

2

)

60

.

1

(

60

]

)

17

.

2

(

)

95

.

0

(

)

68

.

1

[(

20

´

-

+

+

=

08

.

15

=

Then the ANOVA table is below :

Analysis of Variance

Source

SS

d.f.

MS

F

Shelves

15.08

2

7.54

5.21

Error

82.47

57

1.447

Total

97.55

59

Since

15

.

3

05

.

0

,

57

.

2

=

f

F

f

, there do appear to be significant differences among the shelves in terms of fiber content, as well as sugar content.

(c) The grocery store strategy is to place high sugar/low fiber cereals at the eye height of school children where they can easily see them.

12.2

(a) The boxplot indicates that the number of finger taps increases with higher doses of caffeine.

(b)

Analysis of Variance

Source

SS

d.f.

MS

F

Dose

61.400

2

30.700

6.181

Error

134.100

27

4.967

Total

195.500

29

Since

35

.

3

05

.

0

,

27

.

2

=

f

F

f

, there do appear to be significant differences in the numbers of finger taps for different doses of caffeine.

(c) From the plot of the residuals against the predicted values, the constant variance assumption appears satisfied. From the normal plot of the residuals, the residuals appear to follow the normal distribution.

12.5

(a) The boxplot indicates that HBSC has the highest average hemoglobin level, followed by HBS, and then HBSS.

(b)

Analysis of Variance

Source

SS

d.f.

MS

F

Disease type

99.889

2

49.945

49.999

Error

37.959

38

0.999

Total

137.848

40

Since

23

.

3

05

.

0

,

38

.

2

=

f

F

f

, there do appear to be significant differences in the hemoglobin levels between patients with different types of sickle cell disease.

(c) From the plot of the residuals against the predicted values, the constant variance assumption appears satisfied. From the normal plot of the residuals, the residuals appear to follow the normal distribution.

12.8

Sugar : The number of comparisons is

3

2

3

2

=

÷

÷

ø

ö

ç

ç

è

æ

=

÷

÷

ø

ö

ç

ç

è

æ

a

,

Then

468

.

2

0083

.

0

,

57

3

2

05

.

0

,

57

=

=

´

t

t

,

and the Bonferroni critical value is

56

.

1

20

2

999

.

1

468

.

2

2

0083

.

0

,

57

=

´

=

×

×

n

s

t

.

For the Tukey method,

40

.

3

05

.

0

,

57

,

3

»

q

,

And the Tukey critical value is

52

.

1

20

1

999

.

1

40

.

3

1

05

.

0

,

57

,

3

=

´

=

×

n

s

q

Comparison

j

i

y

y

-

Significant?

Bonferroni

Tukey

1 vs. 2

5.05

Yes

Yes

1 vs. 3

1.30

No

No

2 vs. 3

3.75

Yes

Yes

Fiber : The Bonferroni critical value is

939

.

0

20

2

203

.

1

468

.

2

=

´

And the Tukey critical value is

915

.

0

20

1

203

.

1

40

.

3

=

´

Comparison

j

i

y

y

-

Significant?

Bonferroni

Tukey

1 vs. 2

0.73

No

No

1 vs. 3

0.49

No

No

2 vs. 3

1.22

Yes

Yes

12.10

The number of comparisons is

3

2

3

2

=

÷

÷

ø

ö

ç

ç

è

æ

=

÷

÷

ø

ö

ç

ç

è

æ

a

,

Then

136

.

3

0017

.

0

,

38

3

2

01

.

0

,

38

=

=

´

t

t

,

And, since

1

999

.

0

=

=

s

, the form of the Bonferroni confidence interval is

j

i

j

i

n

n

y

y

1

1

)

1

)(

136

.

3

(

+

±

-

For the Tukey method,

104

.

3

2

39

.

4

2

01

.

0

,

38

,

3

=

»

q

,

And the form of the Tukey confidence interval is

j

i

j

i

n

n

y

y

1

1

)

1

)(

104

.

3

(

+

±

-

For the LSD method,

712

.

2

2

01

.

0

,

38

=

t

, and the form of the LSD confidence interval is

j

i

j

i

n

n

y

y

1

1

)

1

)(

712

.

2

(

+

±

-

The 99% confidence intervals are summarized in the table below :

Comparison

j

i

y

y

-

Bonferroni

Tukey

LSD

Lower

Upper

Lower

Upper

Lower

Upper

HBSC(3) vs. HBS(2)

1.670

0.392

2.948

0.406

2.934

0.564

2.776

HBS(2) vs. HBSS(1)

1.918

0.656

3.179

0.669

3.166

0.825

3.010

HBSC(3) vs. HBSS(1)

3.588

2.462

4.713

2.475

4.700

2.614

4.561

Since all of these intervals are entirely above 0, all of the types of disease have significantly different hemoglobin levels from one another. Note that the Bonferroni method has the widest intervals, followed by the Tukey. The LSD intervals are narrowest because there is no adjustment for multiplicity.

12.33

(a) Note that

2

1

2

2

1

1

n

n

y

n

y

n

y

+

+

=

Therefore,

2

2

2

2

1

1

)

(

)

(

y

y

n

y

y

n

MSA

SSA

-

+

-

=

=

)

1

1

(

)

(

)

(

)

(

)

(

)

)(

(

)

(

)

(

)

(

)

(

)

(

)

(

2

1

2

2

1

2

1

2

2

1

2

1

2

2

1

2

2

1

2

1

2

1

2

2

1

2

2

1

2

1

2

2

2

1

2

2

1

2

2

1

2

2

1

2

1

1

2

2

2

1

2

1

2

1

2

2

1

2

2

1

1

2

2

2

2

1

2

2

1

1

1

1

n

n

y

y

n

n

y

y

n

n

n

n

y

y

n

n

n

n

n

n

y

y

n

n

n

n

y

y

n

n

n

n

y

y

n

n

n

n

y

y

n

n

n

n

y

n

y

n

y

n

n

n

y

n

y

n

y

n

+

-

=

+

-

=

+

-

+

=

+

-

+

+

-

=

ú

û

ù

ê

ë

é

+

-

+

ú

û

ù

ê

ë

é

+

-

=

ú

û

ù

ê

ë

é

+

+

-

+

ú

û

ù

ê

ë

é

+

+

-

=

(b)

2

2

1

2

2

2

1

)

1

1

(

)

(

t

n

n

s

y

y

MSE

MSA

F

=

+

-

=

=

(c)

2

,

,

,

1

,

,

1

a

n

a

n

a

n

t

t

f

F

f

F

f

f

f

=

+

+

«

***The following is a past exam problem with solutions using SAS.

Fifteen subjects were randomly assigned to three treatment groups X, Y and Z (with 5 subjects per treatment). Each of the three groups has received a different method of speed-reading instruction. A reading test is given, and the number of words per minute is recorded for each subject. The following data are collected:

X

Y

Z

700

480

500

850

460

550

820

500

480

640

570

600

920

580

610

(c) Are these treatments equally effective? Test at α = 0.05.

(d) If these treatments are not equally good, please use Tukey’s procedure with α = 0.05 to make pairwise comparisons.

(e) Please write a SAS program to answer the above two questions.

Solution : This is one-way ANOVA with 3 samples and 5 observations per sample.

(a) The ANOVA Procedure

Dependent Variable: WORDS

Sum of

Source DF Squares Mean Square F Value Pr > F

Model 2 215613.3333 107806.6667 16.78 0.0003

Error 12 77080.0000 6423.3333

Corrected Total 14 292693.3333

The p-value of the ANOVA F-test is 0.0003, less than the significance level α = 0.05. Thus we conclude that the three reading methods are not equally good.

(b)

The ANOVA Procedure

Tukey's Studentized Range (HSD) Test for WORDS

NOTE: This test controls the Type I experimentwise error rate, but it generally has a higher Type

II error rate than REGWQ.

Alpha 0.05

Error Degrees of Freedom 12

Error Mean Square 6423.333

Critical Value of Studentized Range 3.77278

Minimum Significant Difference 135.22

Means with the same letter are not significantly different.

Tukey Grouping Mean N GROUP

A 786.00 5 X

B 548.00 5 Z

B

B 518.00 5 Y

Thus we found that method X is superior to methods Y and Z, while methods Y and Z are not significantly different at the family-wise error rate of 0.05 using Tukey’s Studentized Range test.

(c) The SAS code is as follows:

DATA READING;

INPUT GROUP $ WORDS @@;

DATALINES;

X 700 X 850 X 820 X 640 X 920 Y 480 Y 460 Y 500

Y 570 Y 580 Z 500 Z 550 Z 480 Z 600 Z 610

;

RUN ;

PROC ANOVA DATA=READING ;

TITLE ‘Analysis of Reading Data’ ;

CLASS GROUP;

MODEL WORDS = GROUP;

MEANS GROUP / TUKEY;

RUN;

_1226947776.unknown
_1226950310.unknown
_1257718629.unknown
_1257718934.unknown
_1257719883.unknown
_1257721015.unknown
_1257721225.unknown
_1257721531.unknown
_1257722005.unknown
_1257722502.unknown
_1257722587.unknown
_1257722421.unknown
_1257721939.unknown
_1257721386.unknown
_1257721469.unknown
_1257721284.unknown
_1257721133.unknown
_1257721180.unknown
_1257721104.unknown
_1257720474.unknown
_1257720643.unknown
_1257720980.unknown
_1257720681.unknown
_1257720883.unknown
_1257720540.unknown
_1257720336.unknown
_1257720446.unknown
_1257720163.unknown
_1257718990.unknown
_1257719829.unknown
_1257719849.unknown
_1257719039.unknown
_1257719774.unknown
_1257718773.unknown
_1257718910.unknown
_1257718797.unknown
_1257718719.unknown
_1257718743.unknown
_1257718681.unknown
_1227865391.unknown
_1257718419.unknown
_1257718550.unknown
_1257718598.unknown
_1257718491.unknown
_1257718517.unknown
_1257718469.unknown
_1257092195.unknown
_1257092196.unknown
_1257092095.unknown
_1257092194.unknown
_1226950986.unknown
_1226951054.unknown
_1226951209.unknown
_1227013491.unknown
_1226951218.unknown
_1226951208.unknown
_1226951021.unknown
_1226950709.unknown
_1226950805.unknown
_1226950525.unknown
_1226948294.unknown
_1226948744.unknown
_1226950215.unknown
_1226950309.unknown
_1226950163.unknown
_1226948544.unknown
_1226948553.unknown
_1226948506.unknown
_1226947963.unknown
_1226948165.unknown
_1226948182.unknown
_1226948120.unknown
_1226947837.unknown
_1226947880.unknown
_1226947777.unknown
_1226946680.unknown
_1226947362.unknown
_1226947724.unknown
_1226947774.unknown
_1226947775.unknown
_1226947753.unknown
_1226947773.unknown
_1226947744.unknown
_1226947634.unknown
_1226947635.unknown
_1226947467.unknown
_1226947183.unknown
_1226947325.unknown
_1226947327.unknown
_1226947324.unknown
_1226946925.unknown
_1226947182.unknown
_1226946902.unknown
_1226945653.unknown
_1226945927.unknown
_1226946071.unknown
_1226946072.unknown
_1226946034.unknown
_1226945915.unknown
_1226945916.unknown
_1226945726.unknown
_1226944823.unknown
_1226945138.unknown
_1226945264.unknown
_1226944882.unknown
_1226944698.unknown
_1226944796.unknown
_1226944378.unknown
_1226944697.unknown