Growing interest in the development of measures to ensure health care (policy and delivery) is better informed by the results of relevant and reliable research (Evidence Based Medicine).
’the integration of individual expertise with the best available external evidence from systematic research’.
failure to combine evidence to address research questions: possibly imprecise; prone to selection bias.
first step in the chain by which research evidence can inform policy and practice.
formal quantitative process for combining the evidence about a treatment effect.
‘the statistical analysis of a large collection of analysis results from individual studies for the purpose of integrating the findings’. (Glass, 1976)
Quantitative: analytical or statistical part of a systematic review.
Purpose: assess combined evidence when several clinical
trials have been conducted (replication of same basic experiment).
Recent: Meta-anlysis in health services and clinical
trials research 80’s onwards Peto, Yusuf; Chalmers, I; Chalmers, T.
Advantage: more precise estimate of treatment effect than
in single studies more powerful: can estimate small, but
clinically relevant effects, with high degree of precision.
Problems: accounting for study heterogeneity; publication bias.
Observational: Meta analyses are observational as opposed to
experimental studies!
Planning: vital stage of the review process and the rules of the procedure should be formalised a priori in a study protocol (objectives, research questions, methods. General guidelines (Armitage et al. 2002)
Data from the different trials should be kept separate so that treatment contrasts derived from individual trials are pooled rather than the original data. Pooling the original data reduces precision and may cause bias.
Inclusion criteria should be clear - differences inevitable; treatments, patient characteristics, outcomes etc. should be ’similar’.
Trials should be randomised trials with adequately ’blind’ assessment. Protocol departures should be handled in same way preferably ITT.
All relevant data should be included: unpublished studies!
Different approaches different results.
Point estimates from different studies will almost certainly differ.
results vary due to sampling error (random variation) as opposed to systematic differences in estimates: underlying true effect the same in each study.
Random variation can be handled using so called fixed effects methodology.
variability exceeds that expected due to sampling differences:
’real’ differences in estimated effects.
Note heterogeneity can occur when effects are in the same direction or different directions.
Examine heterogeneity and present results.
Plot of effect size and corresponding confidence interval for each study on a single
axis.
The pooled estimate and confidence interval is also displayed.
Size of the plotting symbol is often proportional to the reciprocal of the variance. More precise estimates more influence.
plot (z-score) versus reciprocal of the standard error: .
Unnumbered Figure: Link
: treatment effects the same in all primary studies:
() versus
not all effect sizes are equal.
Test statistic:
where is the estimated treatment effect in study .
is the weighted
estimator of the treatment effect and is the weight
for study with denoting the variance of the estimate from
study .
Distribution of Q: approximately on degrees
of freedom under
Reject if Q is significantly large.
Limitations: lacks power, large sample sizes can result in
rejecting null when difference small.
Each study estimate is given a weight inversely proportional to its variance: .
For each of the studies let denote the estimate of the treatment effect and the variance of the estimate (e.g. log odds ratio/difference in treatment group means.
all effects equal versus not all equal.
where .
NOTE Formula for depends on measure of effect (in notes previously)!
Example: Treatment of common cold; binary outcome variable.
Example: Effectiveness of Antibiotics for the Common Cold (Sutton et al, 2000, Sec. 4.3.1).
Endpoint: cure or general improvement within first 7 days?
Binary outcome: yes/no.
Data from five randomised clinical trials: antibiotics
vs. placebo
Study | No. of patients | No. cure | ||
---|---|---|---|---|
antibiotics | control | antibiotics | control | |
1 | 154 | 155 | 67 | 69 |
2 | 146 | 142 | 46 | 48 |
3 | 174 | 87 | 166 | 83 |
4 | 13 | 16 | 9 | 10 |
5 | 129 | 59 | 117 | 56 |
Parameter of interest is the log-odds-ratio.
Can represent study specific data in standard tabular form:
Treatment Group | ||||
---|---|---|---|---|
1 | 2 | Total | ||
Outcome | Yes | |||
observed | No | |||
Total |
Maximum likelihood estimation:
Study | |||
---|---|---|---|
1 | -0.041 | 19.033 | -0.779 |
2 | -0.104 | 15.820 | -1.653 |
3 | 0.000 | 2.544 | 0.000 |
4 | 0.301 | 1.593 | 0.478 |
5 | -0.650 | 2.257 | -1.466 |
=41.257 | = -3.430 |
Combined estimate of the log-odds-ratio:
Variance of pooled estimate:
Approximate confidence interval for the log-odds ratio:
Antilog (0.679, 1.248)
Mantel-Haenszel Method for Combining Odds Ratios (Sutton et al, 2000, Sec. 4.3.1).
History: method first described for use in stratified case-control studies.
tables from studies; table from study with patients:
Treatment | Failure | Success |
---|---|---|
Experimental | ||
Control |
Example: Effectiveness of Antibiotics for the Common Cold (Sutton et al, 2000, Sec. 4.3.1).
Endpoint: cure or general improvement within first 7 days.
Data from five trials: antibiotics vs. placebo
Study | No. of patients | No. cure | ||
---|---|---|---|---|
antibiotics | control | antibiotics | control | |
1 | 154 | 155 | 67 | 69 |
2 | 146 | 142 | 46 | 48 |
3 | 174 | 87 | 166 | 83 |
4 | 13 | 16 | 9 | 10 |
5 | 129 | 59 | 117 | 56 |
pooled estimate of the OR:
95% confidence interval (normal approximation): [0.70;
1.29]
Conclusion?
To avoid bias we need to ensure that relevant primary
studies included.
Extensive literature searching (including grey matter) may not produce unbiased sample.
Plot of sample size (or reciprocal of standard error) versus treatment effect.
small true effect - small effect size - not significant - may not be published.
small true effect - small effect size - significant- likely to be published.
large effect size, more likely to be significant - more likely to be published.
Result: lack of ‘small effect size’ small studies in funnel plot skewed. larger effects in smaller studies; smaller effects larger studies.
Unnumbered Figure: Link
Comments?
Altman DG (1991) Practical statistics for medical research. Chapman & Hall, London.
Bock J (1998) Bestimmung des Stichprobenumfangs. Oldenbourg Verlag, Muenchen.
Jones B, Kenward MG (1990) Design and analysis of cross-over trials. Chapman & Hall, London.
Piantadosi S (1997) Clinical trials: A methodologic perspective. Wiley, New York.
Senn S (1993) Cross-over trials in clinical research. Wiley, Chichester.
Senn S (1997) Statistical issues in drug development. Wiley, Chichester.
Sutton AJ et al (2000) Methods for meta-analysis in medical research. Wiley, Chichester.