lund 2009

42
Biostatistical analysis some basic issues Jonas Ranstam PhD

Upload: jonas-ranstam

Post on 29-May-2015

169 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Lund 2009

Biostatistical analysissome basic issues

Jonas Ranstam PhD

Page 2: Lund 2009

Plan

Statistical principles and their consequences when writing manuscripts for publication in scientific journals

How does this information relate to your research?

General discussion

Page 3: Lund 2009

Statistics is not mathematics

Mathematics is about deduction; statistics starts with data.

John Nelder

Page 4: Lund 2009
Page 5: Lund 2009

Litmus test

A simple test for the acidity of a substance.

Student's t-test

A simple test for the significance of a finding.

Page 6: Lund 2009

Litmus test

A simple test for the acidity of a substance.

Student's t-test

A simple test for the significance of a finding.

Page 7: Lund 2009

Differences between litmus tests and statistical tests

Concentrated hydrochloric acid always turns blue litmus paper red.

A clinically significant difference in systolic blood pressure is not always statistically significant.

Page 8: Lund 2009

Differences between litmus tests and statistical tests

Whenever a blue litmus paper remains blue it has not been exposed to concentrated hydrochloric acid (evidence of absence).

A statistically insignificant difference in systolic blood pressure may be clinically significant (absence of evidence).

Page 9: Lund 2009

Differences between litmus tests and statistical tests

The interpretation of a blue litmus paper turning red is independent of the number of tests performed (no multiplicity issues).

The interpretation of a statistical test showing significance depends on the number of tests performed (multiplicity issues).

Page 10: Lund 2009

Differences between litmus tests and statistical tests

A sample is litmus tested to know more about the sample.

A sample is statistically tested to know more about the population from which the sample is drawn.

Page 11: Lund 2009

A population of black and white dots

Population types

Finite (having 100 dots)

Superpopulation (infinite, but symbolized by the 100 dots)

Page 12: Lund 2009

Sampling dots from the population

Page 13: Lund 2009

?

Sampling uncertainty

What is the sampling uncertainty?

The sample is usually obvious, but what is the population?

What do the sampled items represent, a finite or a super-population?

Page 14: Lund 2009

Sampling uncertainty

The Central Limit Theorem (CLT) states:

The mean of the sampling distribution of means equals the mean of the population from which the samples were drawn.

The variance of the sampling distribution of means equals the variance of the population from which the samples were drawn divided by the size of the samples.

The sampling distribution of means will approximate a Gaussian distribution as sample size increases.

yi ~ N(μ

p; σ

p/√n)

Page 15: Lund 2009

Measurement uncertainty

A measurement, Yi, has two components: the measured

objects true measure, μ0, and a measurement error, e

i:

Yi = μ

0 + e

i

A measurement error is typically unknown, but the population of errors may have a known a distribution:

ei ~ N(μ

e; σ

e)

If μe ≠ 0 measurements will be biased and the greater σ

e the

lower the measurement's precision.

Page 16: Lund 2009

Combining sampling uncertainty and measurement errors

As long as measurements are not biased, measurement errors reduce statistical precision.

This may, however, in turn lead to (dilution) bias or "attenuation by errors"

Page 17: Lund 2009

Alt. 1. Statistical hypothesis testing

Sampling uncertainty described by a probability

H0: π = π

0

H1: π ≠ π

0

p = Prob(drawing a sample looking like H1 | H

0)

If unlikely reject H0 (“statistically significant”)

Page 18: Lund 2009

Alt. 2. Confidence interval

Sampling uncertainty described by a range of values likely (usually 95%) to include an estimated parameter

πE (π

L, π

U)

Page 19: Lund 2009

Effect

Statistically significant effect

Inconclusive

Statistically significant reversed effect

p < 0.05

n.s.

p < 0.05

P-value Conclusion from confidence intervals

P-values vs. confidence intervals

0

Page 20: Lund 2009

0Effect

Clinically significant effect

Statistically and clinically significant effect

Statistically, but not necessarily clinically, significant effect

Inconclusive

Neither statistically nor clinically significant effect

Statistically significant reversed effect

p < 0.05

p < 0.05

n.s.

n.s.

p < 0.05

P-value Conclusion from confidence intervals

P-values vs. confidence intervals

Statistically, but clinically insignificant effect p < 0.05

Page 21: Lund 2009

ICMJE Uniform requirements for manuscripts...

Results

“When possible, quantify findings and present them with appropriate indicators of measurement error or uncertainty (such as confidence intervals).”

“Avoid relying solely on statistical hypothesis testing, such as the use of P values, which fails to convey important information about effect size.”

Page 22: Lund 2009

Present observed data using

a) central tendency (mean, median or mode), b) variability (SD or range) and c) number of observations (n).

Describe parameter estimates

a) with 95% confidence intervals (2SEM) and b) number of observations in the sample (n).

Page 23: Lund 2009

Many scientists do not understand the difference between sample and population

+/-SD or +/-SEM?

SD is a measure of variability.

SEM is a measure sampling uncertainty.

Note

+/- 1SEM corresponds to a 68% confidence interval, +/- 2SEM corresponds to a 95% confidence interval.

Page 24: Lund 2009

ICMJE Uniform requirements for manuscripts...

Methods

Describe statistical methods with enough detail to enable a knowledgeable reader with access to the original data to verify the reported results.

This part of the manuscript could be written prior to the experiment.

Page 25: Lund 2009

Sample size calculation

An essential part of designing a study, or experiment, is planning for uncertainty.

Both too much and too little uncertainty in results are unethical, because research resources are scarce and should be used rationally.

Page 26: Lund 2009

Sample size calculation

Sampling uncertainty increases with the variability of the studied variable.

Sampling uncertainty decreases with the number of independent observations in the sample.

Smaller sampling uncertainty is required to detect small differences than large ones.

Page 27: Lund 2009

Example 1. A vaccine trial

Without protection 30% will fall ill.

Investigating a protective effect, with 5% false positive and 20% false negative error rate requires.

Protection Nr of patients 90% 72 80% 94 70% 128 60% 180 50% 268 40% 428

Page 28: Lund 2009

Example 2. Side effect surveillance

Guillain-Barrés syndrome: Incidence = 1x10-5 personyrs.

To investigate the side effect with 5% false positive and 20% false negative error rate requires.

Risk increase Number of patients100 times 1 098 50 times 2 606 20 times 9 075 10 times 26 366 5 times 92 248 2 times 992 360

Page 29: Lund 2009

Multiplicity

Each tested null hypothesis have, with 5% significance level, 5% chance of being false positive.

Testing n null hypotheses at 5% significance level leads to an overall chance of at least one false positive test of 1 - 0.95n

n overall p1 0.0502 0.0983 0.1434 0.1865 0.226

Page 30: Lund 2009

Multiplicity

Each tested null hypothesis have, with 5% significance level, 5% chance of being false positive.

Testing n null hypotheses at 5% significance level leads to an overall chance of at least one false positive test of 1 - 0.95n

n overall p1 0.0502 0.0983 0.1434 0.1865 0.226

Some experiments have multiplicity problems severe enough to make conventional statistical testing meaningless.

Page 31: Lund 2009

Sampling from a population without variation

Page 32: Lund 2009

Sampling from a population without variation

Without variation there is no sampling uncertainty to evaluate.

Do not test deterministic outcomes, like sex distribution after matching for sex or baseline imbalance after randomisation.

Page 33: Lund 2009

How to report laboratory experiments

Guidelines to promote the transparency of research has been developed in clinical medicine and epidemiology: CONSORT, STROBE, PRISMA, etc.

No such guidelines for reporting laboratory experiments.

Page 34: Lund 2009

Would reporting guidelines be useful?

A systematic review (Roberts et al. BMJ 2002;324:474-476) shows that the reporting generally is inadequate: only 2 of 44 papers described the allocation of analysis units.

Page 35: Lund 2009

General principles

The experiment should be described in a way that makes it possible for the reader to repeat the experiment.

The statistical analysis should be described with enough detail to allow a reader with access to original data to verify reported results.

Page 36: Lund 2009

Introduction section

What is the purpose of the experiment?

What hypotheses will you test (in general terms)?

Page 37: Lund 2009

Material and methods section

What is the design of your experiment?

What is your sample and population?

What is your analysis unit?

Page 38: Lund 2009

Statistics section

What statistical methods did you use?

Did you check whether the assumptions were fulfilled?

How did you do that?

Were the assumptions fulfilled?

Page 39: Lund 2009

Results section (observation)

What was the mean or median value?

What was the variation (SD or range)?

How many observations did you have?

Page 40: Lund 2009

Results section (inference)

What hypotheses did you test (specifically)?

What was the effect size (parameter estimate)?

What was its 95% confidence intervals?

If you present p-values:

Present the actual p-value, e.g. p = 0.45 or p = 0.003, unless p < 0.0001

Do NOT write p > 0.05 or ns or p = 0.0000 or stars

Page 41: Lund 2009

Discussion section

What is your strategy for multiplicity?

What are your interpretation of presented p-values?

Can you see any problems with bias?

What are your overall conclusion?

Page 42: Lund 2009

Discussion1. Describe an experiment you have performed or plan to perform and state the purpose of this.

2. Present the sample and the population from which the sample is drawn.

3. Suggest how the sampling uncertainty could be presented in a manuscript submitted for publication in a scientific journal using p-values and confidence intervals.

HelpRanstam J. Sampling uncertainty in medical research. Osteoarthritis Cartilage 2009;17:1416-1419.

Ranstam J. Reporting laboratory experiments. Osteoarthritis Cartilage 2009 (in press) doi:10.1016/j.joca.2009.07.006.