inferential statistics introduction. if both variables are categorical, build tables... convention:...

Post on 18-Jan-2016

219 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Inferential Statistics

Introduction

If both variables are categorical, build tables . . .

Convention: Each value of the independent (causal) variable has its own column

One table: insert the observed frequencies (# cases that share corresponding values of the ind. and dep. variables)

Another table: insert %. Compute separately for each value of the independent varaible (columns total 100%)

Bivariate (two variables)

Additional independent (control) variables: Construct a first-order partial table for each level of the new variable

Multivariate (three or more variables)

Job Stress

Position on police force

Low

High

Total

Sergeant

86

30

116

Patrol Officer

24

60

84 Total 110 90 N= 200

Job Stress

Position on police force

Low

High

Sergeant

78%

33%

Patrol Officer

22%

67%

Total 100% 100%

r = .87

r2 = .76

Hypothesis: POVERTY CRIME

For additional (control) variables, usepartial correlation or multiple regression

If both variables are continuous, use the r (correlation) and r2 (regression) statistics

Bivariate (two variables)

Multivariate (three or more variables)

Inferential statistics• Go beyond simple evaluations, such as comparing proportions or

calculating a correlation (r)

• Inferential statistics allow us to legitimately “generalize” our findings – apply the results from a sample to a population

• Since we are projecting our findings to a population, we must draw samples. If our sample is a population, we cannot use these methods.

• We must use probability sampling (i.e., random sampling)

• We calculate a “test statistic”, such as an r, X2, t, F, Beta, etc. If this statistic is sufficiently large we can say that there is a “statistically significant” relationship between variables.

How inferential statistics work

Always based on a POPULATION from which a probability (preferably – random) sample is taken

Excepting Chi-Square, which has its own “sampling” distribution, it is assumed that the dependent variable scores in the population are normally distributed around a mean or centerpoint

This centerpoint is the score on the dependent variable that one would expect by randomly drawing a single case from the population

To assess the effect of independent variables, the computer analyzes the dependent variable scores for a particular sample. For example, in the difference between means test, the centerpoint is the average difference between all possible sample means. (This is automatically computed.) If the difference between a sample mean and the centerpoint is large enough to wind up in the distribution’s “tail” (shaded area), the relationship is deemed statistically “significant”. Essentially the same logic applies to all inferential statistics.

Systematic and Error Effects

• Whenever we “observe” that one variable may be causing corresponding changes in another variable, we are really seeing the sum of two things:

– A “systematic” effect, meaning that portion of the relationship that is produced by the action of the independent variable

– An “error” or “chance” effect, meaning that portion that is produced by random factors

• Imagine that the entire surface area of this shape represents the observed relationship between two variables

• How does removing what is due to chance affect our estimate of the strength of the relationship between the variables?

• Imagine that our calculations determine that this portion of the relationship is due to chance

• The null hypothesis holds that the relationship between variables (including any difference between groups) is caused by chance. Essentially, that our working hypothesis is incorrect.

• To “reject” the null hypothesis we must demonstrate that the association between variables is substantially larger than would be produced by chance alone.

• The first step is to compute a “test statistic” – an r, X2, t, F, Beta, etc.

• Based on the size of the test statistic, the sample size, etc. the computer calculates the probability that the null hypothesis is true.

• If at least ONE asterisk appears next to a test statistic (r, t, F, X2, Beta, etc.), the statistic is sufficiently large to overcome the null hypothesis.

[test statistic]* = .05 (5 chances in 100 that the null hypo. is true)[test statistic]** = .01 (1 chance in 100 the null hypo. is true)[test statistic]*** = .001 (1 chance in 1,000 the null hypo is true)

Null Hypothesis

Tests of Significance• The specific test to be used depends on the level of measurement of the variables

• All variables categorical: Chi-Square (X2)• Independent variable categorical, dependent variable continuous: T-test and ANOVA• All variables continuous: Correlation and regression

• Tests of significance yield a statistic, which is the ratio of:

Systematic effect The “real” influence of the independent variables

STATISTIC =(e.g., r, Chi-Square, Chance/random effect “t”, z, “F”, Beta) The apparent influence, actually produced by chance

• The LARGER the numerator, the smaller the denominator, the larger the statistic

• As statistics get bigger, the probability that the null hypothesis is true falls

• When the probability that the null hypothesis is true falls to five chances in hundred (.05) or less, the relationship between the variables is deemed “statistically significant”

Estimating the error betweensample and population statistics

• Standard error of the mean

– An estimate of the amount of the (chance) error we can expect between the means of repeated samples and the population “parameter”

s (standard deviation of our sample)Sx = _____ ___ n-1 (sq. root of sample size minus 1)

– We can’t use s without adjustment because the standard deviation of a sample contains real (systematic) differences in addition to differences that are due to error

• Confidence limit: The range of values within which we are confident the population mean will fall

• Center point for estimating this range is the mean of a single, randomly drawn sample

cl =x z (Sx)

• How “confident” must we be? In social science research, we do not want to exceed a chance of more than 5 in 100 of being wrong

– z-scores represent an actual value on the variable being measured – the more extreme the value, the more extreme the z-score

– Cases with extreme values are so rare that we could say they do not “belong” to the population from which the sample was drawn

Confidence Limit and z scores

95% of the area under a standard normal curve is in the range 1.96z to +1.96z. So, for z, we usually use 1.96

Since the means of repeated samples are normally distributed around the population mean, we can use the normal curve and z scores to find the confidence limits

Distribution of sentences for known population of inmates

24

38

82

38

18

0

10

20

30

40

50

60

70

80

90

1 2 3 4 5

N = 200 inmates, mean sentence 2.94 yrs.

Sentence in years

Class exercise – confidence limit

• Using data from the offender sample drawn earlier, calculate a confidence limit into which the population parameter will fall 95 out of 100 times

s Sx = _____ ___ n-1

cl =x z (Sx)

Police recruit IQ test• Sample of 100 recruits tested for IQ (it is well established that IQ scores are normally distributed in the general population).

x = 120 n = 100 s = 16

• Standard error of the mean: s Sx = --------- ___ n-1

• Confidence limit:

cl =x z (Sx)

There is a 95% probability that the mean IQ of the population from which this sample was drawn falls between these scores:

_______ and ________

lower limit upper limit

x = 120• n = 100• s = 16

• s 16 Sx = _____ = ________ = 1.61 ___ ____ n-1 99

• cl = 120 1.96 (1.61)

• 95% probability that, based on the statistics of this sample, the mean IQ of the population from which the sample was drawn falls between 116.84 and 123.16

top related