Bench-to-bedside review: Avoiding pitfalls in critical care meta-analysis – funnel plots, risk estimates, types of heterogeneity, baseline risk and the ecologic fallacy
Critical Care volume 12, Article number: 220 (2008)
Meta-analysis can be a powerful tool for demonstrating the applicability of a concept beyond the context of individual clinical trials and observational studies, including exploration of effects across different subgroups. Meta-analysis avoids Simpson's paradox, in which a consistent effect in constituent trials is reversed when results are simply pooled. Meta-analysis in critical care medicine is made more complicated, however, by the heterogeneous nature of critically ill patients and the contexts within which they are treated. Failure to properly adjust for this heterogeneity risks missing important subgroup effects in, for example, the interaction of treatment with varying levels of baseline risk. When subgroups are defined by characteristics that vary within constituent trials (such as age) rather than features constant within each trial (such as drug dose), there is the additional risk of incorrect conclusions due to the ecological fallacy. The present review explains these problems and the strategies by which they are overcome.
Meta-analysis is a tool for quantitative systematic review of observational studies and controlled trials that weights available evidence based on the numbers of patients included, the effect size, and often statistical tests of agreement with other trials. Meta-analysis may be particularly suited to critical care medicine. Trials in intensive care typically enrol patients with a variety of pathologies, which can make demonstrating treatment efficacy difficult. These trials are usually underpowered for subgroup analyses. Multicentre trials can increase power with more patients, but between-centre heterogeneity can limit this benefit. Although between-centre heterogeneity can be accounted for, statistical techniques are evolving and imperfect . Conducting a trial in a single centre removes between-centre heterogeneity, but when such trials (for example, those of early goal-directed therapy for severe sepsis  and of tight glycaemic control in critically ill patients ) find treatment effects, physicians can be reluctant to implement the findings if they suspect they were unique to the study institution [4, 5]. The ability to quantitatively detect subgroup effects within heterogeneous populations and to demonstrate external validity should make meta-analyses fundamental components of the critical care literature.
Unfortunately, meta-analysis in critical care can be misleading. A 1998 meta-analysis found albumin use in critically ill patients associated with a 6% increase in absolute mortality . A 6,997-patient randomised controlled trial could not confirm this finding . Meta-analyses do not always agree, but even high-quality reviews attempting to reconcile their differences – such as the review that demonstrated the superiority of sucralfate over histamine receptor-2 antagonists  – have been contradicted later by definitive clinical trials .
Patients in critical care trials are often studied solely because of their presence in an intensive care unit, sometimes even before their diagnosis is known. Interventions can be delivered with differing fidelity and have effects dependent on the baseline risk, which is seldom constant across trials . Other aspects of what constitutes intensive care are frequently highly variable. If not appropriately addressed, heterogeneity in patients, interventions and context can produce misleading conclusions. The present review highlights how these conclusions arise, and explores approaches to these problems.
Identifying publication bias: inspecting the funnel plot
Conducting a meta-analysis using published study results has the advantage of using evidence already subjected to peer scrutiny. Positive studies have a greater chance of being published, however, which can falsely exaggerate the results . To combat publication bias, meta-analyses may also include unpublished data , a process facilitated by the modern requirement to prospectively register clinical trials.
Regardless of whether they incorporate unpublished data, all meta-analyses should check for inclusion bias. A funnel plot (Figure 1) graphs each study's estimated treatment effect against an estimate of the precision of this estimate , such as the standard error or the number of patients included. Less precise estimates should become increasingly spread out, forming a funnel (Figure 1a). Assymmetry suggests omission of some studies, although study heterogeneity and the small study effect  can produce the same pattern (Figure 1b, c). Publication bias can also produce a hollow plot, where studies showing minimal effect are missing while strongly positive or negative studies are included (Figure 1e). Publication bias can be statistically tested and even corrected , but the suggested methods have not gained widespread acceptance .
Publication bias may be a particular problem in critical care medicine. Only 49 out of 139 critical care meta-analyses adequately reported an absence of publication bias . Inspecting the funnel plot, however, is effective only when there are enough trials . In critical care, with comparatively few trials for any clinical question but with a relatively open research community, the certainty that all appropriate trials have been included is at least as useful.
Combining different measures of treatment effect: risk difference, risk ratio or odds ratio?
A drug that reduces mortality by 20% will save a greater number of people when used in a high-risk population versus a low-risk population. Displaying the range of measures of absolute effect (usually the risk difference; Table 1) in a meta-analysis demonstrates this fact, making the data intuitively accessible. If attempting to combine estimates of treatment effect in different populations, however, it is better to use measures of relative risk – such as odds ratios, risk ratios or hazard ratios [17, 18].
Unfortunately the choice of a suitable summary statistic is more complicated than would appear. Odds ratios and risk ratios both reflect relative risk, but although related they are fundamentally different (Table 1). The odds ratio is the only valid measure of association in case–control studies. As meta-analysis developed from the statistical approaches used to combine case–control studies, the odds ratio has become the default measure of effect . Although theoretically the mathematically superior approach, in certain circumstances there are empirical reasons for challenging this default. For example, a review of 551 meta-analyses found median study heterogeneity was lower when a pooled relative risk (rather than odds ratio) was used . In individual studies, however, many analyses were clearly better performed with one or the other, while some analyses showed no difference. The factors influencing which summary statistic was preferable could not be predicted, and neither statistic was found clearly superior overall.
Meta-analysis avoids Simpson's paradox
It might be tempting to simply pool data from all patients in a number of seemingly similar trials as if they came from one large study. Unfortunately, this can lead to Simpson's paradox , where (for example) a beneficial effect in each study can become an apparently detrimental effect when the data are aggregated. In practice, Simpson's paradox only arises when pooling the results of observational studies in which there is a severe imbalance in an important confounding factor along with unequal group assignments. Randomisation in controlled trials should prevent this occurrence.
Meta-analysis overcomes Simpson's paradox by accounting for the enrolment of patients in different studies. There is a published example , but the three theoretical studies presented in Table 2 may be clearer. In this case, all three studies find treatment associated with lower mortality than the placebo. If the data are simply pooled, the relative risk is reversed. A fixed-effects meta-analysis finds the risk ratio to be 0.898, however, which agrees with the individual study conclusions. Figure 2 shows a graphical representation of a similar effect.
Are studies too different to combine? Detecting heterogeneity using the Q test and the I2 statistic
It is not surprising that a new drug tested at a moderate dose in very ill patients receiving excellent care might have a different effect when tested at a high dose, with inconsistent delivery and monitoring, in patients across a spectrum of disease severity in a hospital struggling to provide basic services. Some trials are simply too different to combine. Heterogeneity amongst patients, contexts or interventions is problematic only if the intervention's efficacy is influenced by one of these factors. In critical care, this is often the case: for example, with activated protein C, which was effective in a mixed population containing high-risk patients  but not in a low-risk population .
If the confidence intervals of the Forest plot (Figure 3) do not overlap, the true treatment effect in those studies is probably different. Heterogeneity can also be statistically tested, most commonly using the Q test. A small P value means the null hypothesis (of study homogeneity) should be rejected – and the studies should not be combined, at least not without adjustment. Unfortunately, this test's power is relatively low when there are few studies, but is prone to overdetect heterogeneity when there are many studies. Threshold P values are arbitrarily often set higher or lower than 0.05.
A more recent approach is to report the I2 statistic , which quantifies the percentage of total variation between studies that is due to heterogeneity rather than due to chance. A value of 0% indicates no heterogeneity, with the scale increasing to 100%. In contrast to the Q test, the I2 statistic facilitates the comparison of meta-analyses of different sizes. There is no value of I2that is considered too high: the original description suggested I2 values of 25%, 50% and 75% indicated low, moderate and high levels of heterogeneity, respectively, but that other factors (such as consistency of direction of effect and the clinical characteristics of the study) precluded definition of an arbitrary threshold.
Dealing with heterogeneity: study selection – fixed-effects models and random-effects models
If heterogeneity is detected, it is simplest to exclude the outlier studies, preferably with some justification. Excluding studies simply because they do not agree with the majority defeats the purpose of the meta-analysis. To counter the suspicion that inclusion criteria have been adjusted to achieve a desired effect, an analysis plan to deal with heterogeneity should be specified in advance. If heterogeneity is detected, the plan should identify which trial characteristics (such as quality, drug dose, baseline risk, and so forth) will be grounds for exclusion.
Heterogeneity is sometimes more informative than problematic. The cleanest signal would be found by looking only at trials with similar eligibility and exclusion criteria, drug doses, and hospital contexts. Showing a qualitatively consistent treatment effect despite significant heterogeneity, however, is equivalent to showing the treatment works in a variety of contexts – the definition of external validity.
Rather than exclude studies, the other simple approach is to statistically allow for differences between trials using a random-effects model. The intuitive assumption underlying meta-analysis is that of fixed effects: that a number of different studies are being combined to estimate one true effect of the intervention. In contrast, the random-effects model does not assume an intervention has the same effect in each of the studies. An individual study is therefore considered a random sample from a hypothetical population of similar studies. The study's effect estimate is consequently considered less precise.
Informative confounding: the two types of heterogeneity
Rather than adjust for different effects in different studies, it can be better to realise that doing so obscures potentially important information. If a drug is of benefit in men but does some harm to women, knowing this might be more important than using a single summary statistic to conclude overall moderate advantage.
Between-study variation in the constituent trials of a meta-analyses falls into two categories: trial-level factors and patient-level factors. Trial-level factors apply to all patients in each trial: for example, drug dose. In contrast, patient-level factors such as age and sex vary both within and between trials. Heterogeneous trial-level factors are generally easily taken into consideration, but patient-level confounding can be particularly difficult.
Dealing with trial-level heterogeneity: subgroup meta-analysis and meta-regression
The two approaches to dealing with trial-level heterogeneity are subgroup meta-analysis and meta-regression. Consider a number of trials of a drug, some of which use the oral route and some the intravenous route. Conducting separate meta-analyses on each of these strata makes clinical sense, particularly if there is a difference in the observed effect between the two routes. One could then conclude for each route whether the drug was beneficial. This analysis will also indirectly estimate the effect of the administration route, but the possibility of confounding (for example, perhaps patients who could use the oral route were less unwell) makes such a comparison unwise.
The alternative approach is meta-regression , a statistical model quantifying the effect of various study characteristics on the estimated overall effect. This approach is particularly useful for understanding the effect of a factor present at more than two levels, the classic example being drug dose. Meta-regression will quantify how much of the between-trial heterogeneity is explained by the various drug doses used, and is most useful when there is a significant treatment effect, a large number of studies, sufficient between-study variation in the postulated confounding variable, and sufficient heterogeneity among the treatment effects .
The main criticism of these techniques is that they constitute data-dredging. If trials are split into too many subgroups or too many factors are incorporated into meta-regressions, the probability of a false-positive conclusion due to multiple comparisons increases. Conversely, the small number of trials on which most of these analyses are based means power is limited, so true associations may be missed . As with subgroup analyses in clinical trials, splitting a meta-analysis into subgroups is considered by many to be hypothesis generating at best. Nonetheless, the hypotheses may be stronger if the subgroups are based on prerandomisation characteristics, were planned a priori and allowed an adequately powered analysis, and if there is statistical adjustment for multiple testing .
Dealing with patient-level heterogeneity: avoiding the ecological fallacy
Using meta-regression to account for differences in the types of patients enrolled is possible, but potentially problematic. Such an analysis must use average patient characteristics. The relationship between the effect estimate and average patient characteristics across trials may not be the same as that relationship within trials, as is demonstrated in Figure 4. In the upper part of the figure, treatment effect is related to age within each trial, but is not related to the mean age across trials. In the lower part of the figure, the opposite is true: there is a relationship across trials, but not within trials. This is a classic example of the ecological fallacy, in which incorrect inferences about individual characteristics are made based upon aggregate statistics.
The term ecological fallacy was coined as an explanation for a phenomenon observed in the 1930 US census . Literacy had been positively correlated with immigrant numbers in each US state, which lead to the unlikely conclusion that immigrants were more literate. When literacy within each state was examined, however, the opposite relationship was observed. The explanation was that immigrants tended to settle in states where the native population was more literate.
The only way to avoid the ecological fallacy when considering possible patient-level confounding in a meta-analysis is to examine data from individuals.
Baseline risk: an example of patient-level heterogeneity particularly relevant to critical care
It is simplest to assume the treatment in question has the same relative effect in patient groups with different baseline risks. This assumption can be inappropriate, particularly in intensive care, where treatments (such as activated protein C) often have substantial potential for harm as well as for benefit. If a treatment effect is heavily influenced by baseline risk (for example, harmful in low-risk patients and beneficial in those at high risk), it is necessary to adjust for risk in a multivariable model or stratified analysis. Failure to do so gives the appearance of random variation (or the variation might be misassigned to another factor, such as trial quality), whereas in reality it is an important finding. The traditional approach has been to use the event rate in the control group as a surrogate for baseline risk. This approach introduces bias due to regression to the mean, and is now considered inappropriate . Alternatives have been proposed  but, as with other patient-level factors, the ideal solution is to investigate the interaction of treatment effect with individual patient characteristics , ideally using individual patient data.
Individual patient data meta-analysis
Analysing data at an individual patient level is the most powerful meta-analytic technique available. Statistical significance is crudely determined by a ratio of explained variation over unexplained variation. The ability to account for individual patient covariates, for treatment differences between studies, and for the interactions of these factors means a greater proportion of the unexplained variation can be accounted for – increasing the power of the meta-analysis.
Other advantages of individual patient data meta-analyses over those analyses using aggregate patient data include the ability to undertake sufficiently powered exploratory subgroup analyses whilst avoiding the ecologic fallacy, to adjust for differences in baseline risk, to analyse time to event data rather than single-point outcome statistics, to update survival information, to carry out a detailed check of the primary data, and to reanalyse the data using potentially more appropriate methods . Examples of such re-analyses include the ability to check the statistical assumptions of regression models, to reanalyse the data using intention-to-treat analysis, and to include patients inappropriately excluded from the original analysis.
A simulation study comparing aggregate data meta-regression and individual patient data meta-analysis found the individual patient data approach had higher statistical power. There was little agreement between the estimates of effect size between the two methods . Meta-analysis of individual patient data 'is acknowledged as the gold standard' . Nonetheless, individual patient data meta-analyses are performed 20 times less frequently than those using aggregate patient data [34, 35], because access to detailed trial results is difficult.
Most meta-analyses are conducted retrospectively, when a series of smaller trials have failed to demonstrate a convincing result due to lack of power, or are conducted to explore subgroup effects. Where possible, the Cochrane Collaboration advocates prospective meta-analysis , which overcomes inconsistencies in data collection, entry criteria, study protocols and outcome measures, as well as the criticism of data-dredging. Even if individual studies are adequately powered for their primary endpoints, this is unlikely to be true for secondary outcomes and for important subgroups. Additionally, trials are conventionally funded to have 80% power, implying a 20% chance of missing a true treatment effect. Power calculations are often based on poor-quality data, and tend to be overoptimistic – such as in the recent trial of activated factor VII that hypothesised a 33% relative improvement in outcome for patients with acute intracerebral haemorrhage . These factors all argue for consideration of prospective meta-analysis in the planning of any clinical trial.
Reporting of meta-analysis
Even if appropriately conducted, a meta-analysis must be adequately reported to facilitate scrutiny of the results. Unfortunately in critical care medicine this is frequently not the case. A systematic review of 139 meta-analyses relevant to critical care found overall quality poor , with the most common omissions being failure to report whether a comprehensive literature search was conducted, how inclusion bias was addressed, and assessment of the validity of the included studies. The mean quality improved after publication of the Quality of Reporting of Meta-analyses (QUORUM) guidelines .
Better awareness of the issues surrounding meta-analysis particularly relevant to critical care – especially the existence of the ecological fallacy and the possible interaction of treatment with baseline risk – will hopefully improve the performance, reporting and critical review of this valuable technique. Many pitfalls are avoided if a meta-analysis uses individual patient data and is prospectively planned, suggesting future clinical investigators should carefully consider the advantages and disadvantages of this approach.
This article is part of a review series on Translational research, edited by John Kellum.
Other articles in the series can be found online at http://ccforum.com/articles/theme-series.asp?series=CC_Trans
Komarek A, Lesaffre E, Legrand C: Baseline and treatment effect heterogeneity for survival times between centers using a random effects accelerated failure time model with flexible error distribution. Stat Med. 2007, 26: 5457-5472. 10.1002/sim.3083.
Rivers E, Nguyen B, Havstad S, Ressler J, Muzzin A, Knoblich B, Peterson E, Tomlanovich M: Early goal-directed therapy in the treatment of severe sepsis and septic shock. N Engl J Med. 2001, 345: 1368-1377. 10.1056/NEJMoa010307.
van den BG, Wouters P, Weekers F, Verwaest C, Bruyninckx F, Schetz M, Vlasselaers D, Ferdinande P, Lauwers P, Bouillon R: Intensive insulin therapy in the critically ill patients. N Engl J Med. 2001, 345: 1359-1367. 10.1056/NEJMoa011300.
Peake S, Webb S, Delaney A: Early goal-directed therapy of septic shock: we honestly remain skeptical. Crit Care Med. 2007, 35: 994-995. 10.1097/01.CCM.0000257481.37623.3B.
Angus DC, Abraham E: Intensive insulin therapy in critical illness. Am J Respir Crit Care Med. 2005, 172: 1358-1359. 10.1164/rccm.2508009.
Bunn F, Lefebvre C, Li L, Po ALW, Roberts I, Schierhout G: Human albumin administration in critically ill patients: systematic review of randomised controlled trials. Cochrane Injuries Group Albumin Reviewers. BMJ. 1998, 317: 235-240.
Finfer S, Bellomo R, Boyce N, French J, Myburgh J, Norton R: A comparison of albumin and saline for fluid resuscitation in the intensive care unit. N Engl J Med. 2004, 350: 2247-2256. 10.1056/NEJMoa040232.
Cook DJ, Reeve BK, Guyatt GH, Heyland DK, Griffith LE, Buckingham L, Tryba M: Stress ulcer prophylaxis in critically ill patients. Resolving discordant meta-analyses. JAMA. 1996, 275: 308-314. 10.1001/jama.275.4.308.
Cook D, Guyatt G, Marshall J, Leasa D, Fuller H, Hall R, Peters S, Rutledge F, Griffith L, McLellan A, Wood G, Kirby A: A comparison of sucralfate and ranitidine for the prevention of upper gastrointestinal bleeding in patients requiring mechanical ventilation. Canadian Critical Care Trials Group. N Engl J Med. 1998, 338: 791-797. 10.1056/NEJM199803193381203.
Peelen L, de Keizer NF, Peek N, Bosman RJ, Scheffer GJ, de JE: Influence of entry criteria on mortality risk and number of eligible patients in recent studies on severe sepsis. Crit Care Med. 2005, 33: 2178-2183. 10.1097/01.CCM.0000181733.16353.2C.
Easterbrook PJ, Berlin JA, Gopalan R, Matthews DR: Publication bias in clinical research. Lancet. 1991, 337: 867-872. 10.1016/0140-6736(91)90201-Y.
McAuley L, Pham B, Tugwell P, Moher D: Does the inclusion of grey literature influence estimates of intervention effectiveness reported in meta-analyses?. Lancet. 2000, 356: 1228-1231. 10.1016/S0140-6736(00)02786-0.
Egger M, Davey SG, Schneider M, Minder C: Bias in meta-analysis detected by a simple, graphical test. BMJ. 1997, 315: 629-634.
Sterne JA, Egger M, Smith GD: Systematic reviews in health care: investigating and dealing with publication and other biases in meta-analysis. BMJ. 2001, 323: 101-105. 10.1136/bmj.323.7304.101.
Alderson P, Green S: Cochrane Collaboration Open Learning Material for Reviewers. 2002, Oxford, UK: Cochrane Collaboration, [http://www.cochrane-net.org/openlearning/index.htm]
Delaney A, Bagshaw SM, Ferland A, Manns B, Laupland KB, Doig CJ: A systematic evaluation of the quality of meta-analyses in the critical care literature. Crit Care. 2005, 9: R575-R582. 10.1186/cc3803.
Deeks JJ: Issues in the selection of a summary statistic for meta-analysis of clinical trials with binary outcomes. Stat Med. 2002, 21: 1575-1600. 10.1002/sim.1188.
Egger M, Smith GD, Phillips AN: Meta-analysis: principles and procedures. BMJ. 1997, 315: 1533-1537.
Simpson EH: The interpretation of interaction in contingency tables. J R Stat Soc. 1951, 13: 238-241.
Hanley JA, Theriault G: Simpson's paradox in meta-analysis. Epidemiology. 2000, 11: 613-614. 10.1097/00001648-200009000-00022.
Bernard GR, Vincent JL, Laterre PF, LaRosa SP, Dhainaut JF, Lopez-Rodriguez A, Steingrub JS, Garber GE, Helterbrand JD, Ely EW: Efficacy and safety of recombinant human activated protein C for severe sepsis. N Engl J Med. 2001, 344: 699-709. 10.1056/NEJM200103083441001.
Abraham E, Laterre PF, Garg R, Levy H, Talwar D, Trzaskoma BL, Francois B, Guy JS, Bruckmann M, Rea-Neto A, Rossaint R, Perrotin D, Sablotzki A, Arkins N, Uttebback BG, Macias WL: Drotrecogin alfa (activated) for adults with severe sepsis and a low risk of death. N Engl J Med. 2005, 353: 1332-1341. 10.1056/NEJMoa050935.
Higgins JP, Thompson SG, Deeks JJ, Altman DG: Measuring inconsistency in meta-analyses. BMJ. 2003, 327: 557-560. 10.1136/bmj.327.7414.557.
Thompson SG, Higgins JP: How should meta-regression analyses be undertaken and interpreted?. Stat Med. 2002, 21: 1559-1573. 10.1002/sim.1187.
Schmid CH, Stark PC, Berlin JA, Landais P, Lau J: Meta-regression detected associations between heterogeneous treatment effects and study-level, but not patient-level, factors. J Clin Epidemiol. 2004, 57: 683-697. 10.1016/j.jclinepi.2003.12.001.
Geddes J: Meta-analysis in the assessment of treatment outcome. J Psychopharmacol. 2006, 20: 67-71. 10.1177/1359786806066056.
Cook DI, Gebski VJ, Keech AC: Subgroup analysis in clinical trials. Med J Aust. 2004, 180: 289-291.
Robinson WS: Ecological correlations and the behavior of individuals. Am Sociol Rev. 1950, 15: 351-357. 10.2307/2087176.
Sharp SJ, Thompson SG, Altman DG: The relation between treatment benefit and underlying risk in meta-analysis. BMJ. 1996, 313: 735-738.
Sharp SJ, Thompson SG: Analysing the relationship between treatment effect and underlying risk in meta-analysis: comparison and development of approaches. Stat Med. 2000, 19: 3251-3274. 10.1002/1097-0258(20001215)19:23<3251::AID-SIM625>3.0.CO;2-2.
Stewart LA, Clarke MJ: Practical methodology of meta-analyses (overviews) using updated individual patient data. Cochrane Working Group. Stat Med. 1995, 14: 2057-2079. 10.1002/sim.4780141902.
Lambert PC, Sutton AJ, Abrams KR, Jones DR: A comparison of summary patient-level covariates in meta-regression with individual patient data meta-analysis. J Clin Epidemiol. 2002, 55: 86-94. 10.1016/S0895-4356(01)00414-0.
Chalmers I: The Cochrane collaboration: preparing, maintaining, and disseminating systematic reviews of the effects of health care. Ann N Y Acad Sci. 1993, 703: 156-163. 10.1111/j.1749-6632.1993.tb26345.x.
Lyman GH, Kuderer NM: The strengths and limitations of meta-analyses based on aggregate data. BMC Med Res Methodol. 2005, 5: 14-10.1186/1471-2288-5-14.
Simmonds MC, Higgins JP, Stewart LA, Tierney JF, Clarke MJ, Thompson SG: Meta-analysis of individual patient data from randomized trials: a review of methods used in practice. Clin Trials. 2005, 2: 209-217. 10.1191/1740774505cn087oa.
Cochrane Collaboration. [http://www.cochrane.org]
Mayer SA, Brun NC, Begtrup K, Broderick J, Davis S, Diringer MN, Skolnick BE, Steiner T: Efficacy and safety of recombinant activated factor VII for acute intracerebral hemorrhage. N Engl J Med. 2008, 358: 2127-2137. 10.1056/NEJMoa0707534.
Moher D, Cook DJ, Eastwood S, Olkin I, Rennie D, Stroup DF: Improving the quality of reports of meta-analyses of randomised controlled trials: the QUOROM statement. Quality of reporting of meta-analyses. Lancet. 1999, 354: 1896-1900. 10.1016/S0140-6736(99)04149-5.
Education Queensland Exploring Data. [http://exploringdata.cqu.edu.au/sim_par.htm]
Hintze J: NCSS, PASS and GESS [computer program]. 2006, Kaysville, UT: NCSS
The authors declare that they have no competing interests.
About this article
Cite this article
Reade, M.C., Delaney, A., Bailey, M.J. et al. Bench-to-bedside review: Avoiding pitfalls in critical care meta-analysis – funnel plots, risk estimates, types of heterogeneity, baseline risk and the ecologic fallacy. Crit Care 12, 220 (2008). https://doi.org/10.1186/cc6941
- Publication Bias
- Critical Care
- Individual Patient Data
- Baseline Risk
- Ecological Fallacy