i. statistical tests:

21
I. Statistical Tests: A. Why do we use them? Namely: we need to make inferences from incomplete information or uncertainty But we want them to be “educated” or “calculated” guesses We want to make them with some minimal likelihood of a wrong inference (or conversely, a maximal chance of a correct inference). Note: If we had full information, we wouldn’t need statistical tests – in fact, they would be meaningless. B. What do they involve?

Upload: carlow

Post on 01-Feb-2016

31 views

Category:

Documents


0 download

DESCRIPTION

I. Statistical Tests:. Why do we use them? Namely: we need to make inferences from incomplete information or uncertainty But we want them to be “educated” or “calculated” guesses - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: I. Statistical Tests:

I. Statistical Tests:A. Why do we use them?

• Namely: we need to make inferences from incomplete information or uncertainty But we want them to be “educated” or

“calculated” guesses We want to make them with some minimal

likelihood of a wrong inference (or conversely, a maximal chance of a correct inference).

• Note: If we had full information, we wouldn’t need statistical tests – in fact, they would be meaningless.

B. What do they involve?• The “Null Hypothesis Testing” procedure• What is the underlying logic of NHT?

Page 2: I. Statistical Tests:

I. Statistical Tests: (cont.)B. What do they involve?

• The basic logic of testing a “null hypothesis”? Testing our research hypothesis against a null

hypothesis (any differences found = sampling error) Compare sample value against value specified by Null-

Hypothesis (considering sampling variability) Are the difference so large that it is unlikely to be

sampling error (considering sampling variability)?

• The form of the research hypothesis determines the “directionality” of the test Non-directional hypothesis = 2-tailed test Directional hypothesis = 1-tailed test 2-tailed test = more conservative 1-tailed test = more likely to find significant result

Page 3: I. Statistical Tests:

I. Statistical Tests: (cont.)

C. What makes a test result turn out “statistically significant”?

1) Magnitude of the effect or difference

2) Amount of variability in data

3) Sample size

4) Probability level selected (“alpha level”)

5) Directional vs. non-directional test

6) Number of total tests performed

Page 4: I. Statistical Tests:

II. How to do Statistical Tests?A. 3 Basic Tasks: Testing hypotheses about:

1) 1 mean (or 1 sample)

2) 2 means (or 2 samples)

3) 3 or more means (or samples)

1. One Sample tests• Testing a sample statistic against a hypothetical

value• Relevant when we have a specific prediction

about the mean(rarely)• This is the simplest form of statistical test• Use a Z or t test here (whichever is relevant)

Page 5: I. Statistical Tests:

II. How to do Statistical Tests? (cont.)2. Two Sample tests

• Test difference between two means or groups (against Null-H that they are equal)

a) Difference between 2 independent means Compare 2 separate groups on 1 variable

b) Difference between 2 correlated or paired means (also: “matched groups” or “repeated measures”) Compare 2 variables on 1 group

• Use Z-test when population variance is known• Use t-test when population variance is unknown

and must be estimated from the sample• What are the assumptions of these tests?

Page 6: I. Statistical Tests:

Tests of Mean Differences in SPSS

Page 7: I. Statistical Tests:

Test of 1 Mean = “1 Sample T-test”

Page 8: I. Statistical Tests:

Test of 2 Separate Means: “Indep. Samples T-Test”

Page 9: I. Statistical Tests:

Independent Samples T-Test (cont.)

Page 10: I. Statistical Tests:

Test of 2 Correlated Groups = “Paired Samples T-test”

Page 11: I. Statistical Tests:

SPSS T-test Output

Page 12: I. Statistical Tests:

II. How to do Statistical Tests? (cont.)

3. Tests with 3 or groups (& means)?a) Why not calculate multiple pair-wise tests?

• Can become unwieldy if we have no specific predictions about the pattern

• Overall error rate increases rapidly as the number of tests increases Note: per-test vs overall error rates

b) The Solution = Use an overall statistical test of differences among multiple (2 or more) groups This is called “Analysis Of Variance” (or commonly called ANOVA)

Page 13: I. Statistical Tests:

III. What is ANOVA?

1. Similar in logic to t-test comparison of two means except that we use squared deviations rather than simple differences

2. Derive two independent estimates of the population variance for the scores

a. One based on the variations among the group means (between-group estimate)

b. One based on the variations among scores in each of the groups (within-group estimates) (pooled)

Each group is taken as a random sample from the same population (under the null hypothesis)

Page 14: I. Statistical Tests:

C. What is ANOVA? (cont.)

3. Compare the two separate variance estimates (between-group vs. within-group)

Compute a ratio of the two estimates:(between means) / (within groups)

If the variation in the group means is larger than the variations within the separate groups, then the groups are really different (i.e., they are not from the same population)

4. Use a statistic that is computed as the ratio of two variances called the F statistic

has a calculated probability distribution

Page 15: I. Statistical Tests:

C. What is ANOVA? (cont.)

2. F = (Variance of Means)/(Variance of scores)

– F-distribution depends on two parameters:– These represent the “degrees of freedom” in the

two different variance estimates in the ratio• df1 = degrees of freedom in numerator (group means)

k - 1• df2 = degrees of freedom in individual scores

N - k

Page 16: I. Statistical Tests:

C. What is ANOVA? (cont.)

3. The logic of the F-test?3. If the F-ratio is much larger than 1.0, which means much

larger differences between group means than expected from random sampling errors

4. Then we reject the null hypothesis of equality of means

5. This decision is based on the probability distribution of the F-statistic so that Type 1 error is .05 or less.

• The null hypothesis and the alternative hypothesis of the ANOVA F-test?

1 2 3 i

i j

H0:

Hr: (for at least 2 groups)

Page 17: I. Statistical Tests:

C. What is ANOVA? (cont.)

5. What does the F-test tell us?– The F-test is a non-directional, omnibus test– The F-test doesn’t tell us which two means are

different or in what direction– It simply affirms that we can say that “at least 2

group means are different” (with a .95 confidence level)

6. What if we want to be more specific?a) Can use inspection-by-eyeball method

b) Can use post-hoc comparisons (or contrasts)

Page 18: I. Statistical Tests:

C. The Logic of ANOVA (continued)

6. “Post Hoc Comparisons” identify which pairs of means are significantly different, while

– Control overall error rate of the whole set of comparisons– Adjust each individual pairwise comparison to keep the

overall error rate at desired level (e.g., .05)

7. Which “post hoc comparison” procedure to use?– Many different procedures (and frameworks) have been

developed and are available in SPSS– In this class, select only one: Tukey’s HSD procedure

(in SPSS, is simply labeled Tukey)

Page 19: I. Statistical Tests:
Page 20: I. Statistical Tests:
Page 21: I. Statistical Tests:

Descriptives

HOW OFTEN R ATTENDS RELIGIOUS SERVICES

370 3.20 2.464 .128

535 3.59 2.651 .115

482 4.29 2.635 .120

1387 3.73 2.632 .071

Liberal

Moderate

Conservative

Total

N Mean Std. Deviation Std. Error

ANOVA

HOW OFTEN R ATTENDS RELIGIOUS SERVICES

267.632 2 133.816 19.843 .000

9333.521 1384 6.744

9601.152 1386

Between Groups

Within Groups

Total

Sum ofSquares df Mean Square F Sig.

Multiple Comparisons

Dependent Variable: HOW OFTEN R ATTENDS RELIGIOUS SERVICES

Tukey HSD

-.395 .176 .063 -.81 .02

-1.095* .179 .000 -1.52 -.67

.395 .176 .063 -.02 .81

-.700* .163 .000 -1.08 -.32

1.095* .179 .000 .67 1.52

.700* .163 .000 .32 1.08

(J) Conservative-Liberal(3-category)Moderate

Conservative

Liberal

Conservative

Liberal

Moderate

(I) Conservative-Liberal(3-category)Liberal

Moderate

Conservative

MeanDifference

(I-J) Std. Error Sig. Lower Bound Upper Bound

95% Confidence Interval

The mean difference is significant at the .05 level.*.

Post Hoc Tests