MetaReview

Free Meta-Analysis Calculator Online: Compute Effect Sizes & Heterogeneity

Calculate pooled effect sizes, heterogeneity statistics, and generate forest plots in minutes. No coding, no installation, no cost. A complete guide with formulas, real examples, and step-by-step instructions.

Table of Contents

  1. What Is a Meta-Analysis Calculator?
  2. Types of Effect Sizes You Can Calculate
  3. Understanding Heterogeneity Statistics
  4. Fixed-Effect vs Random-Effects Models
  5. Step-by-Step: Run a Meta-Analysis with MetaReview
  6. Meta-Analysis Calculator Comparison Table
  7. Common Calculation Pitfalls
  8. Reporting Meta-Analysis Results
  9. Frequently Asked Questions

What Is a Meta-Analysis Calculator?

A meta-analysis calculator is a computational tool that takes individual study data -- effect sizes, sample sizes, variances, or raw outcome counts -- and produces a pooled (combined) estimate of the overall effect across all included studies. It automates the statistical machinery that underpins meta-analysis: weighting studies by their precision, choosing between fixed-effect and random-effects models, computing heterogeneity statistics, and generating confidence intervals for the pooled result.

Meta-analysis is the quantitative heart of a systematic review. While the systematic review process involves formulating a question, searching databases, screening studies, extracting data, and assessing risk of bias, the meta-analysis step is where the actual numbers are crunched. A meta-analysis calculator handles this computational work, freeing researchers to focus on the decisions that require human judgment: which studies to include, which effect measure to use, how to handle clinical heterogeneity, and how to interpret the results in context.

Why Researchers Need Automated Computation

The mathematics of meta-analysis, while conceptually straightforward, involves dozens of intermediate calculations that are tedious and error-prone when done by hand. For a single meta-analysis of 10 studies using a random-effects model, you need to: (1) calculate the effect size and its variance for each study, (2) compute inverse variance weights, (3) calculate the weighted mean, (4) compute Cochran's Q statistic, (5) estimate tau-squared using an iterative method like DerSimonian-Laird or REML, (6) re-calculate weights incorporating tau-squared, (7) compute the new pooled estimate and its standard error, (8) derive the confidence interval and p-value, and (9) calculate I-squared and prediction intervals. Each step feeds into the next, meaning a single arithmetic error propagates through the entire analysis.

Automated calculators eliminate these errors. They also enforce consistency: every study is weighted using the same formula, every confidence interval uses the same method, and every heterogeneity statistic is calculated correctly. This reproducibility is essential for scientific credibility. When you report that you used MetaReview or R's metafor package, readers and reviewers know exactly what computations were performed, which is far more transparent than saying "we calculated pooled estimates in Excel."

Manual vs Automated Calculation: A Practical Comparison

To illustrate why automated tools matter, consider a simple example. Suppose you are pooling five randomized controlled trials that each report the number of patients who experienced a cardiovascular event in the aspirin group versus the placebo group. To compute the pooled odds ratio manually, you would need to:

  1. For each study, construct a 2x2 table and calculate the log odds ratio: ln(OR) = ln(a*d / b*c), where a = events in treatment, b = non-events in treatment, c = events in control, d = non-events in control.
  2. Calculate the variance of each log(OR) using the formula: Var(ln(OR)) = 1/a + 1/b + 1/c + 1/d.
  3. Compute the inverse variance weight for each study: w_i = 1 / Var(ln(OR_i)).
  4. Calculate the fixed-effect pooled log(OR): sum(w_i * ln(OR_i)) / sum(w_i).
  5. Calculate Q = sum(w_i * (ln(OR_i) - pooled_ln(OR))^2).
  6. Derive I-squared = max(0, (Q - df) / Q * 100%), where df = k - 1.
  7. If using random-effects, estimate tau-squared using DerSimonian-Laird: tau^2 = (Q - df) / (sum(w_i) - sum(w_i^2)/sum(w_i)), truncated to 0 if negative.
  8. Re-weight with random-effects weights: w_i* = 1 / (Var(ln(OR_i)) + tau^2).
  9. Compute the random-effects pooled log(OR) and its standard error.
  10. Back-transform by exponentiating to get the pooled OR and its confidence interval.

That is 10 multi-step calculations for just five studies using one effect measure. Now imagine doing this for 25 studies, with subgroup analyses, sensitivity analyses (leave-one-out), and a different effect measure. The manual approach becomes impractical. A meta-analysis calculator performs all of these steps in under a second, with zero chance of arithmetic error.

What a Good Meta-Analysis Calculator Should Provide

Not all meta-analysis tools are created equal. A high-quality calculator should provide:

MetaReview provides all of these features in a free, browser-based interface that requires no installation, no coding, and no account registration. It is designed to be the fastest path from extracted data to publication-ready meta-analysis results.

Key takeaway: A meta-analysis calculator automates the precise, multi-step statistical computations required to pool study results. Manual calculation is error-prone and impractical for modern systematic reviews. Free online tools like MetaReview make rigorous meta-analysis accessible to every researcher, regardless of programming skill or software budget.

Types of Effect Sizes You Can Calculate

The choice of effect size measure is one of the most consequential decisions in a meta-analysis. It determines how individual study results are expressed, how they are combined, and how the pooled result is interpreted. The effect size must match your data type (binary vs continuous), your study design (RCT vs cohort vs case-control), and your clinical question. Below, we cover the five most commonly used effect sizes in health research meta-analyses, with formulas, interpretation guidance, and real-world examples.

1. Odds Ratio (OR)

The odds ratio compares the odds of an event in the intervention group to the odds of the event in the control group. It is the most commonly used effect measure for binary (dichotomous) outcomes in meta-analysis, particularly for case-control studies and situations where the baseline risk varies substantially across studies.

OR = (a / b) / (c / d) = (a * d) / (b * c) Where: a = events in intervention group b = non-events in intervention group c = events in control group d = non-events in control group Standard error of ln(OR) = sqrt(1/a + 1/b + 1/c + 1/d)

Interpretation: An OR of 1.0 means no difference between groups. An OR less than 1.0 means lower odds in the intervention group (the intervention is protective). An OR greater than 1.0 means higher odds in the intervention group (the intervention increases risk). For example, an OR of 0.70 means the odds of the event are 30% lower in the intervention group.

Clinical example -- Aspirin for cardiovascular prevention: Consider a meta-analysis of randomized trials evaluating low-dose aspirin for preventing myocardial infarction. In the Physicians' Health Study, 139 of 11,037 physicians in the aspirin group had a myocardial infarction compared to 239 of 11,034 in the placebo group. The odds ratio for this single study would be (139 * 10,795) / (10,898 * 239) = 0.576, suggesting aspirin reduced the odds of MI by approximately 42%. When pooled across multiple trials, the meta-analysis calculator weights each study by the inverse of its variance and produces an overall OR with a combined confidence interval.

When to use OR: Case-control studies (where RR cannot be directly calculated), rare events (where OR approximates RR), situations where baseline risk differs substantially across studies, and logistic regression outputs. The OR is also mathematically convenient because it is symmetric on the log scale, making it well-suited for meta-analytic pooling.

Caution: When the outcome is common (event rate above 10-20%), the OR will overestimate the relative risk. An OR of 2.0 does not mean "twice the risk" -- it means "twice the odds." For common outcomes, the RR is more intuitive and clinically interpretable. Misinterpreting OR as RR is one of the most frequent errors in medical literature.

2. Risk Ratio (RR)

The risk ratio (also called relative risk) compares the probability of an event in the intervention group to the probability in the control group. It is generally preferred over the OR for cohort studies and RCTs because it is more intuitive for clinicians and patients: "the risk is 1.5 times higher" is easier to communicate than "the odds are 1.5 times higher."

RR = (a / n1) / (c / n2) Where: a = events in intervention group n1 = total participants in intervention group c = events in control group n2 = total participants in control group Standard error of ln(RR) = sqrt(1/a - 1/n1 + 1/c - 1/n2)

Interpretation: An RR of 1.0 indicates no difference. An RR of 0.80 means a 20% reduction in risk with the intervention. An RR of 1.50 means a 50% increase in risk. The number needed to treat (NNT) can be derived from the RR and the baseline risk: NNT = 1 / (baseline_risk * (1 - RR)).

Clinical example -- Antihypertensive treatment for stroke prevention: In a meta-analysis of antihypertensive drug trials, each study reports the number of stroke events in the treatment versus placebo group. If one trial reports 25 strokes among 2,000 treated patients and 40 strokes among 2,000 control patients, the RR = (25/2000) / (40/2000) = 0.625, indicating a 37.5% reduction in stroke risk. The meta-analysis calculator pools these RRs across all trials using inverse variance weighting on the log scale.

When to use RR: RCTs and prospective cohort studies with binary outcomes, situations where you want to communicate effect in terms of "risk" (more intuitive for clinical audiences), and when baseline event rates are moderate to high.

3. Hazard Ratio (HR)

The hazard ratio is the standard effect measure for time-to-event (survival) data. It compares the instantaneous rate of the event in the intervention group to the rate in the control group at any given time point, accounting for censoring (patients who drop out or are lost to follow-up before experiencing the event).

HR is estimated from survival analysis models (Cox proportional hazards). For meta-analysis, you need: ln(HR) and its standard error (SE). If not directly reported, ln(HR) can be estimated from: - Kaplan-Meier curves (using methods by Parmar, Tierney, or Guyot) - O-E statistics and variance: ln(HR) = (O - E) / V - Reported p-values and event counts (using Tierney's methods) SE(ln(HR)) = (ln(upper CI) - ln(lower CI)) / (2 * 1.96)

Interpretation: An HR of 1.0 indicates no difference. An HR of 0.75 means a 25% lower hazard (instantaneous risk) of the event at any time point. Unlike OR and RR, the HR incorporates the time dimension, making it the appropriate measure when the timing of events matters (overall survival, progression-free survival, time to relapse).

Clinical example -- Adjuvant chemotherapy for colon cancer: A meta-analysis of trials evaluating fluorouracil-based chemotherapy after surgical resection for stage III colon cancer pools hazard ratios for overall survival. Each trial reports an HR from a Cox model comparing chemotherapy to observation. An individual trial might report HR = 0.68 (95% CI: 0.50-0.92), meaning chemotherapy reduced the hazard of death by 32%. The meta-analysis calculator combines these on the log(HR) scale.

When to use HR: Oncology trials (overall survival, progression-free survival), cardiovascular trials with time-to-event endpoints, any study reporting survival analysis. The HR is the only appropriate measure when censoring is present and event timing is clinically relevant.

4. Mean Difference (MD)

The mean difference (also called weighted mean difference or WMD) is used for continuous outcomes when all studies measure the outcome on the same scale. It is the simplest continuous effect measure: the difference in mean values between the intervention and control groups.

MD = Mean_intervention - Mean_control Variance(MD) = SD1^2/n1 + SD2^2/n2 Standard error of MD = sqrt(SD1^2/n1 + SD2^2/n2) Where: SD1, SD2 = standard deviations in intervention and control groups n1, n2 = sample sizes in intervention and control groups

Interpretation: The MD is directly interpretable in the original measurement units. An MD of -5.2 mmHg for systolic blood pressure means the intervention group had, on average, 5.2 mmHg lower blood pressure than the control group. This direct interpretability is the main advantage of MD over SMD.

Clinical example -- Antihypertensive drugs for blood pressure reduction: A meta-analysis of ACE inhibitor trials for hypertension pools mean differences in systolic blood pressure reduction. Study A reports a mean reduction of 12.3 mmHg (SD 8.2, n=120) in the treatment group versus 4.1 mmHg (SD 7.9, n=115) in placebo, yielding MD = -8.2 mmHg. Study B reports treatment mean 10.8 mmHg (SD 9.1, n=85) versus placebo mean 3.5 mmHg (SD 8.6, n=82), yielding MD = -7.3 mmHg. The calculator pools these MDs, weighting each by the inverse of its variance, to produce an overall pooled MD with confidence interval.

When to use MD: All studies use the same measurement instrument and scale (e.g., all report HbA1c in %, blood pressure in mmHg, or Beck Depression Inventory scores). MD should be the default for continuous outcomes when scales are identical, because it preserves clinical interpretability.

5. Standardized Mean Difference (SMD / Hedges' g)

The standardized mean difference is used when studies measure the same underlying construct but use different measurement scales. It expresses the difference between groups in standard deviation units, allowing studies with different instruments to be combined.

Cohen's d = (Mean_intervention - Mean_control) / SD_pooled SD_pooled = sqrt(((n1-1)*SD1^2 + (n2-1)*SD2^2) / (n1 + n2 - 2)) Hedges' g = d * (1 - 3/(4*(n1+n2) - 9)) (Small-sample correction factor, also called J) Variance(g) = (n1+n2)/(n1*n2) + g^2/(2*(n1+n2))

Interpretation: The SMD is measured in standard deviation units. Cohen's benchmarks are commonly used: 0.2 = small effect, 0.5 = medium effect, 0.8 = large effect. Hedges' g corrects for the small upward bias in Cohen's d that occurs with small sample sizes. In meta-analysis, Hedges' g is preferred over Cohen's d because of this correction. An SMD of -0.45 means the intervention group scored about half a standard deviation lower than the control group.

Clinical example -- Cognitive behavioral therapy for depression: A meta-analysis evaluates the effect of CBT on depression symptoms. However, different trials use different depression scales: some use the Hamilton Depression Rating Scale (HDRS, range 0-52), others use the Beck Depression Inventory (BDI, range 0-63), and others use the Patient Health Questionnaire (PHQ-9, range 0-27). Because these scales have different ranges and standard deviations, mean differences cannot be pooled directly. The SMD standardizes each study's result into a common unit (standard deviations), enabling valid pooling. If the pooled SMD is -0.62, it indicates a moderate-to-large effect of CBT on depression, equivalent to about 0.62 standard deviations improvement regardless of which scale was used.

When to use SMD: Studies measure the same construct on different scales (e.g., pain measured with VAS in some studies and NRS in others), psychological outcomes measured with different validated instruments, or quality of life measured with different questionnaires. Do NOT use SMD when all studies use the same scale -- use MD instead, because MD preserves clinical interpretability.

Quick Reference: Choosing Your Effect Size

Outcome TypeSame Scale?Study DesignRecommended Measure
Binary (yes/no)N/ACase-controlOdds Ratio (OR)
Binary (yes/no)N/ARCT / CohortRisk Ratio (RR)
Time-to-eventN/ASurvival analysisHazard Ratio (HR)
ContinuousYesAnyMean Difference (MD)
ContinuousNoAnyStandardized Mean Difference (SMD)
MetaReview supports all of these: Select your effect measure when setting up the analysis. MetaReview handles all the underlying calculations -- log transformations, variance estimation, pooling, and back-transformation -- automatically. You just enter the raw data.

Understanding Heterogeneity Statistics

Heterogeneity is the degree to which the true effect sizes vary across the studies included in a meta-analysis. It is arguably the most important concept in meta-analysis after the pooled effect itself, because high heterogeneity means the pooled estimate may not adequately represent any individual study's true effect. A meta-analysis calculator must compute and report heterogeneity statistics so that researchers and readers can assess how consistent the evidence is.

There are two types of heterogeneity. Clinical heterogeneity refers to differences in patient populations, interventions, comparators, outcomes, and settings across studies -- this is assessed qualitatively by the review authors. Statistical heterogeneity refers to variability in effect sizes beyond what is expected from sampling error -- this is what the calculator quantifies. Statistical heterogeneity may arise from clinical heterogeneity, methodological differences, or both.

I-Squared (I²): The Percentage of True Variation

I-squared is the most widely reported heterogeneity statistic. It describes the percentage of total variability in effect estimates that is due to true differences between studies rather than sampling error (chance). Introduced by Higgins and Thompson in 2002, it has become the standard metric because it is easy to interpret and does not depend on the number of studies or the effect size scale.

I² = max(0, (Q - df) / Q) * 100% Where: Q = Cochran's Q statistic (see below) df = k - 1 (degrees of freedom, k = number of studies)

Interpretation thresholds (based on the Cochrane Handbook, Chapter 10):

I² RangeHeterogeneity LevelClinical Implication
0% - 25%LowStudy results are broadly consistent. The pooled estimate is likely a good summary of the overall effect. A fixed-effect model is usually appropriate.
25% - 50%ModerateSome variation exists beyond chance. Consider exploring sources through subgroup analysis. Both models are reasonable; random-effects is safer.
50% - 75%SubstantialMeaningful differences exist across studies. Subgroup or sensitivity analysis is expected. The pooled estimate should be interpreted cautiously. Investigate whether clinical or methodological differences explain the variation.
75% - 100%ConsiderableStudies are measuring fundamentally different effects. Pooling may not be appropriate. Narrative synthesis or separate subgroup analyses may be more informative than a single pooled estimate. The prediction interval will be very wide.
Important caveat: I-squared should not be interpreted in isolation. A meta-analysis of two studies might have I² = 0% simply because there is insufficient power to detect heterogeneity. Conversely, a meta-analysis of 100 large studies might have I² = 40% that is highly statistically significant but clinically trivial if all effect sizes cluster in a narrow, clinically meaningful range. Always consider I² alongside the Q test, tau-squared, the prediction interval, and the visual spread of the forest plot.

Cochran's Q Test: Is There Significant Heterogeneity?

Cochran's Q is a chi-squared test that evaluates the null hypothesis that all studies share the same true effect size. It is computed as the weighted sum of squared deviations of each study's effect from the pooled effect.

Q = SUM(w_i * (theta_i - theta_pooled)^2) Where: w_i = inverse variance weight of study i theta_i = effect size of study i theta_pooled = fixed-effect pooled estimate Q follows a chi-squared distribution with df = k - 1

Interpretation: A significant Q test (p < 0.10 -- a lenient threshold is used because Q has low statistical power, especially with fewer than 10 studies) indicates that the observed variation in effect sizes is greater than expected by chance alone. However, a non-significant Q does not prove homogeneity -- it may simply reflect low power to detect true heterogeneity.

The Q statistic itself is also used in the calculation of I-squared and tau-squared. It is a building block of the heterogeneity assessment, not a standalone verdict. Reporting both Q (with its p-value and degrees of freedom) and I-squared provides complementary information.

Tau-Squared (τ²): The Between-Study Variance

While I-squared tells you the proportion of variation due to heterogeneity, tau-squared tells you the actual magnitude of between-study variance in the true effect sizes. It is expressed in the same squared units as the effect size, making it directly interpretable (though tau, the square root of tau-squared, is more intuitive as a standard deviation).

DerSimonian-Laird estimator: tau^2 = max(0, (Q - df) / (sum(w_i) - sum(w_i^2)/sum(w_i))) Where: Q = Cochran's Q statistic df = k - 1 w_i = fixed-effect inverse variance weights

Interpretation: Tau-squared represents the variance of the distribution of true effects. If tau^2 = 0, there is no between-study variability (the fixed-effect assumption holds). A large tau-squared means the true effects are widely dispersed. For example, if you are pooling odds ratios and tau^2 = 0.15 on the log(OR) scale, then tau = sqrt(0.15) = 0.39, meaning the standard deviation of the distribution of true log(ORs) is 0.39. Approximately 95% of true effects would fall within plus/minus 1.96*0.39 = 0.76 log-OR units of the pooled mean.

Tau-squared is critical because it is used to calculate random-effects weights. In the random-effects model, each study's weight is 1 / (within-study variance + tau^2). This means that when tau^2 is large, the weights become more equal across studies -- large studies lose their dominance because the between-study variability swamps the within-study sampling error.

Prediction Interval: The Clinically Crucial Statistic

The prediction interval is perhaps the most underused and underreported statistic in meta-analysis. While the confidence interval tells you the uncertainty around the mean pooled effect, the prediction interval tells you the range within which the true effect of a new, future study is expected to fall. It incorporates both the uncertainty of the pooled mean and the between-study variability.

95% Prediction Interval = pooled_estimate +/- t(k-2, 0.025) * sqrt(tau^2 + SE_pooled^2) Where: t(k-2, 0.025) = critical value from t-distribution with k-2 degrees of freedom tau^2 = between-study variance SE_pooled = standard error of the pooled estimate k = number of studies

Why it matters clinically: Imagine a meta-analysis finding a pooled OR of 0.60 (95% CI: 0.45-0.80). This looks like a robust protective effect. But if the 95% prediction interval is 0.25-1.45, it means that in a new clinical setting, the true effect could plausibly favor either treatment or control. A clinician deciding whether to implement the intervention in their hospital needs to know this: the average effect is protective, but it might not apply in every setting. The prediction interval captures this practical uncertainty in a way the confidence interval does not.

IntHout et al. (2016) demonstrated that in many published meta-analyses with statistically significant pooled results, the prediction interval crosses the null, fundamentally changing the clinical interpretation. The GRADE framework also considers prediction intervals when evaluating the certainty of evidence.

Putting Heterogeneity Statistics Together

No single heterogeneity statistic tells the full story. Best practice is to report all four metrics together:

MetaReview computes all four: When you run a meta-analysis in MetaReview, the results panel displays I-squared, Q (with p-value and degrees of freedom), tau-squared, and the prediction interval automatically. You do not need to calculate any of these by hand.

Fixed-Effect vs Random-Effects Models

The choice between a fixed-effect model and a random-effects model is one of the most fundamental decisions in meta-analysis. It reflects your assumption about the nature of the true effect sizes across the included studies, and it directly affects the pooled estimate, the confidence interval width, the study weights, and the interpretation of results. Understanding the difference is essential for any researcher using a meta-analysis calculator.

The Fixed-Effect Model

The fixed-effect model (note: "fixed-effect," singular, not "fixed-effects" -- this is the conventional terminology in the Cochrane Handbook) assumes that all studies in the meta-analysis estimate the same true underlying effect size. Any variation in observed effect sizes across studies is attributed entirely to sampling error (within-study random variation). The model treats the true effect as a single, fixed parameter that every study shares.

Under this model, the pooled estimate is calculated using inverse variance weighting:

Pooled estimate (fixed) = SUM(w_i * theta_i) / SUM(w_i) Where: w_i = 1 / Var(theta_i) (inverse of within-study variance) theta_i = effect size of study i SE(pooled) = 1 / sqrt(SUM(w_i)) 95% CI = pooled +/- 1.96 * SE(pooled)

Consequences of the fixed-effect assumption:

When is the fixed-effect model appropriate? In practice, very few meta-analyses genuinely meet the fixed-effect assumption. It is most defensible when: (1) all studies used virtually identical populations, interventions, comparators, outcomes, and settings (e.g., identical drug at identical dose in the same type of patient); (2) you are interested only in the specific set of included studies (not generalizing to other settings); or (3) there are very few studies (k = 2-3) and there is no statistical power to estimate tau-squared reliably.

The Random-Effects Model

The random-effects model assumes that each study estimates a different true effect size, and that these true effects are drawn from a probability distribution (typically assumed to be normal) with mean mu and variance tau-squared. The model acknowledges that studies differ in populations, intervention details, outcome measurement, and other factors that cause the true effect to vary from study to study.

Random-effects model: theta_i = mu + zeta_i + epsilon_i Where: mu = overall mean of the distribution of true effects zeta_i ~ N(0, tau^2) (between-study random effect) epsilon_i ~ N(0, sigma_i^2) (within-study sampling error) Random-effects weights: w_i* = 1 / (sigma_i^2 + tau^2) Pooled estimate (random) = SUM(w_i* * theta_i) / SUM(w_i*) SE(pooled) = 1 / sqrt(SUM(w_i*))

The key difference is the addition of tau-squared in the weight denominator. This has several consequences:

DerSimonian-Laird Estimator

The DerSimonian-Laird (DL) method is the most commonly used approach for estimating tau-squared in random-effects meta-analysis. It is a method-of-moments estimator: it equates the observed Q statistic to its expected value under the random-effects model and solves for tau-squared.

tau^2_DL = max(0, (Q - (k-1)) / C) Where: C = SUM(w_i) - SUM(w_i^2) / SUM(w_i) Q = Cochran's Q statistic k = number of studies w_i = fixed-effect inverse variance weights If Q <= k-1, then tau^2 = 0 (no heterogeneity detected)

The DL estimator is fast, simple, and does not require iteration. It is implemented in MetaReview and is the default in most meta-analysis software. However, it has known limitations: it can underestimate tau-squared when the number of studies is small (k < 10-15), leading to confidence intervals that are too narrow. Alternative estimators include REML (restricted maximum likelihood), PM (Paule-Mandel), and HKSJ (Hartung-Knapp-Sidik-Jonkman) adjustments that produce more conservative inference with few studies.

How Model Choice Affects Results

To make the impact concrete, consider a hypothetical meta-analysis of 8 studies examining a new anticoagulant versus standard treatment for venous thromboembolism prevention:

AspectFixed-Effect ResultRandom-Effects Result
Pooled OR0.580.62
95% CI0.49 - 0.690.44 - 0.87
p-value< 0.00010.006
I-squared58% (moderate-to-substantial)
95% Prediction IntervalN/A (not applicable)0.28 - 1.37
InterpretationStrong, precise protective effectProtective on average, but the effect may not apply in all settings

Notice that: (1) the random-effects CI is wider, appropriately reflecting between-study variability; (2) the random-effects pooled OR is slightly pulled toward the null, because small studies with more extreme results receive relatively more weight; (3) the prediction interval crosses 1.0, suggesting the treatment may not be beneficial in all settings. The fixed-effect model would have you believe the effect is highly precise and consistent. The random-effects model paints a more nuanced and honest picture.

Common Misconceptions

Practical recommendation: For most meta-analyses, the random-effects model is the appropriate default. Clinical heterogeneity is almost always present to some degree. If I-squared is 0%, the results will be identical to fixed-effect anyway. MetaReview lets you toggle between models to compare, but you should pre-specify your primary model in your protocol.

Step-by-Step: Run a Meta-Analysis with MetaReview

This section walks you through the complete process of performing a meta-analysis using MetaReview's free online calculator. From opening the tool to exporting your final report, the entire workflow takes approximately 10 to 15 minutes for a standard meta-analysis with 5 to 20 studies.

1

Open MetaReview and Select Effect Measure

Open MetaReview in any modern web browser (Chrome, Firefox, Safari, or Edge). No account registration, no software installation, and no payment is required. The tool runs entirely in your browser.

Begin by selecting the appropriate effect size measure for your data. This decision should align with your systematic review protocol and the nature of your included studies:

Not sure which to choose? Refer to the effect size selection guide above, or read our detailed effect size selection tutorial.
2

Enter Study Data (Binary or Continuous)

Enter the extracted data for each included study. MetaReview provides a clean data entry interface with real-time validation.

For Binary Outcomes (OR or RR)

FieldDescriptionExample
Study NameFirst author and yearSmith 2019
Events (Intervention)Number of events in treatment group23
Total (Intervention)Total participants in treatment group150
Events (Control)Number of events in control group45
Total (Control)Total participants in control group148

For Continuous Outcomes (MD or SMD)

FieldDescriptionExample
Study NameFirst author and yearJones 2020
Mean (Intervention)Mean value in treatment group-2.4
SD (Intervention)Standard deviation in treatment group1.8
N (Intervention)Sample size of treatment group85
Mean (Control)Mean value in control group-0.6
SD (Control)Standard deviation in control group2.1
N (Control)Sample size of control group82

MetaReview validates your inputs in real time. It will immediately flag: events exceeding the total sample size, negative standard deviations, zero or negative sample sizes, and missing required fields. Fix any errors before proceeding.

Double-check your data: Data entry errors are the most common source of incorrect meta-analysis results. After entering all studies, review the complete data table within MetaReview against your original data extraction spreadsheet. Pay special attention to intervention/control column assignment and decimal placement.
3

Import from CSV/Excel (Optional)

For meta-analyses with many studies, manual data entry is tedious. MetaReview supports CSV file import for faster data loading.

Prepare your data in a spreadsheet (Excel, Google Sheets, or any application that exports CSV) with one row per study and column headers matching MetaReview's expected format. Click the import button, select your file, and MetaReview will automatically map columns and display a preview of the imported data.

Verify the preview carefully: check that column mapping is correct, all studies are present, and no data was truncated or misaligned. Once confirmed, the imported data populates the data entry table and is ready for analysis.

CSV format tip: Use UTF-8 encoding to ensure study author names with accented characters (e.g., Gonzalez, Muller) import correctly. Save as CSV (UTF-8) from Excel or use Google Sheets' native CSV export.
4

Choose the Analysis Model

Select your statistical model for pooling. MetaReview offers two options:

In the vast majority of published meta-analyses, the random-effects model is the appropriate choice because some degree of clinical heterogeneity is almost always present. When I-squared is 0%, both models produce identical results, so there is no penalty for defaulting to random-effects.

Protocol consistency: Your model choice should be pre-specified in your systematic review protocol (e.g., PROSPERO registration). Using the random-effects model as the primary analysis and the fixed-effect model as a sensitivity check is a common and defensible approach.
5

Review Pooled Results and Heterogeneity

MetaReview automatically computes and displays the complete results panel:

Review these results before proceeding to visualization. If I-squared is above 50%, consider whether subgroup analysis or sensitivity analysis is warranted. If the prediction interval crosses the null, note this for your discussion section.

6

Generate Forest Plot

Navigate to the forest plot view. MetaReview renders a publication-ready forest plot showing:

Customize the forest plot as needed: adjust study sorting (by effect size, weight, year, or entry order), toggle display of weight percentages and CI values, modify axis labels, and configure subgroup display if applicable. The plot updates in real time as you make changes.

Subgroup forest plots: If you have defined subgroups in your data (e.g., by study region, dose level, or risk of bias), MetaReview can display subgroup-specific pooled diamonds within the same forest plot, with an overall diamond at the bottom. This is essential for exploring sources of heterogeneity.
7

Check Publication Bias (Funnel Plot + Tests)

Publication bias occurs when studies with non-significant or unfavorable results are less likely to be published, leading to a systematically biased set of included studies. MetaReview provides tools to assess this:

Limitation: Publication bias tests require at least 10 studies for adequate statistical power. With fewer studies, visual inspection of the funnel plot is still informative but formal tests should not be over-interpreted. Funnel plot asymmetry can also result from genuine heterogeneity (e.g., if larger studies systematically differ from smaller studies in population or intervention), not just publication bias.
8

Export Report (HTML/DOCX/JSON)

Export your complete meta-analysis results in the format you need:

MetaReview generates a results paragraph that you can use as a starting point for your manuscript's methods and results sections. This paragraph follows standard reporting conventions and includes all the key statistics (pooled effect, CI, p-value, I-squared, model used, number of studies) that journal reviewers expect to see.

Complete in 10-15 minutes: From opening MetaReview to downloading your final forest plot and results summary, the entire process takes 10 to 15 minutes for a typical meta-analysis with 5 to 20 studies. This includes data entry, model selection, result review, and export.

Meta-Analysis Calculator Comparison Table

Researchers have several options for computing meta-analyses. The best choice depends on your technical skill level, budget, analytical complexity, and workflow preferences. Below is a detailed comparison of the most widely used meta-analysis calculators and software packages.

Feature MetaReview R / metafor RevMan (Cochrane) Stata (meta) CMA OpenMeta[Analyst] Meta-Mar
Cost Free Free (open source) Free (Cochrane account) $295-$895/year $195-$1,395 Free (open source) Free
Platform Browser (any OS) Desktop (any OS) Desktop + Web Desktop (Win/Mac) Desktop (Windows) Desktop (any OS) Browser
Coding Required No Yes (R) No Yes (Stata syntax) No No No
Installation None R + packages Software download Software + license Software + license Java required None
Effect Sizes: OR Yes Yes Yes Yes Yes Yes Yes
Effect Sizes: RR Yes Yes Yes Yes Yes Yes Yes
Effect Sizes: MD Yes Yes Yes Yes Yes Yes Yes
Effect Sizes: SMD Yes Yes Yes Yes Yes Yes Limited
Effect Sizes: HR Yes Yes Yes (generic inverse variance) Yes Yes Limited No
Forest Plot Yes (SVG/PNG) Yes (highly customizable) Yes Yes Yes Yes Yes (basic)
Funnel Plot Yes Yes Yes Yes Yes Yes No
Subgroup Analysis Yes Yes Yes Yes Yes Yes No
Meta-Regression Planned Yes (full) No Yes Yes Yes No
Sensitivity Analysis Yes (leave-one-out) Yes (all types) Limited Yes Yes Yes No
Publication Bias Tests Egger's, trim-and-fill All (Egger, Begg, trim-fill, PET-PEESE, etc.) Funnel plot only All major tests All major tests Egger's, Begg's No
GRADE Support Planned No (use GRADEpro) Integrated (GRADEpro link) No No No No
PRISMA Support Checklist export No Integrated No No No No
Export Formats SVG, PNG, HTML PDF, SVG, PNG, TIFF, EPS Image, PDF, RevMan format PDF, PNG, EPS Image, CMA format Image, CSV Image
Learning Curve 5-10 minutes Days to weeks 1-2 hours Days to weeks 1-2 hours 30-60 minutes 5 minutes
Citable in Papers Yes Yes (Viechtbauer 2010) Yes (Cochrane standard) Yes Yes (Borenstein et al.) Yes (Wallace et al.) Limited
Best For Beginners, fast results, no-code workflow Power users, custom analyses, meta-regression Cochrane systematic reviews Academic statisticians, epidemiologists Researchers wanting GUI with advanced features Open-source advocates, basic analyses Quick, simple calculations

MetaReview: The Fastest Free Online Calculator

MetaReview occupies a unique position: it combines the ease of use of a simple online calculator with the analytical depth of professional software. It runs in any browser, requires no installation or coding, supports all standard effect sizes and models, and produces publication-ready forest plots. For the majority of systematic reviews published in biomedical journals -- those requiring standard pairwise meta-analysis with OR, RR, MD, or SMD -- MetaReview provides everything you need at zero cost.

R / metafor: The Gold Standard for Power Users

Wolfgang Viechtbauer's metafor package for R is the most comprehensive meta-analysis tool available. It supports every effect size measure, every heterogeneity estimator (DL, REML, PM, ML, HKSJ, and more), multivariate meta-analysis, network meta-analysis (with extensions), meta-regression with multiple moderators, and virtually unlimited plot customization. The trade-off is a steep learning curve: you must be comfortable writing R code, managing packages, and debugging scripts. If you need meta-regression, dose-response analysis, or non-standard effect sizes, metafor is the tool to use.

RevMan: The Cochrane Ecosystem

RevMan (Review Manager) is developed and maintained by the Cochrane Collaboration. It is the standard tool for Cochrane systematic reviews and integrates with Cochrane's risk-of-bias tool and GRADEpro for evidence quality assessment. RevMan Web is gradually replacing the desktop version. For Cochrane-affiliated reviews, RevMan is the expected choice. For non-Cochrane reviews, it offers less flexibility than MetaReview or R.

Stata and CMA: For Institutional Users

Stata (with the built-in meta command in version 16+ or the community-contributed metan) and Comprehensive Meta-Analysis (CMA) are widely used in universities and research institutions that hold site licenses. Both are capable tools with good documentation. Their main limitation is cost: Stata licenses range from $295 to $895 per year, and CMA starts at $195. For individual researchers, graduate students, or researchers in resource-limited settings, the cost is a significant barrier.

Our recommendation: Start with MetaReview for standard meta-analyses. It covers OR, RR, MD, SMD, both models, forest plots, funnel plots, and sensitivity analysis -- enough for 90% of published systematic reviews. Graduate to R/metafor only when you need meta-regression, network meta-analysis, or highly custom analyses. Avoid paying for software when free tools meet your needs.

Common Calculation Pitfalls

Even with automated calculators, certain methodological decisions and data issues can lead to incorrect or misleading meta-analysis results. A good calculator prevents arithmetic errors, but it cannot prevent conceptual errors. Understanding these common pitfalls will help you avoid them and improve the credibility of your review.

1. Zero Cells in 2x2 Tables

A zero cell occurs when one or more cells in a 2x2 table have zero events. For example, if no patients in the treatment group experienced the adverse event, the cell count for "events in treatment" is zero. This creates a mathematical problem: the odds ratio involves division, and a zero in the denominator produces an undefined value. The log(OR) formula requires taking the natural log of zero, which is negative infinity.

How calculators handle it: The standard approach is to add a continuity correction -- typically 0.5 to all four cells of the affected study. This allows the OR and its variance to be computed. However, this correction introduces a small bias, and different correction values (0.5 vs 0.01 vs treatment-arm-specific corrections) can produce different results. Some approaches (the Peto method, or exact methods) avoid the continuity correction entirely.

What you should do: Report whether any studies had zero cells, which continuity correction was applied, and consider a sensitivity analysis using different approaches (e.g., Peto OR for rare events, or excluding studies with zero cells). If many studies have zero cells, it often indicates a very rare event, and the Peto method or exact methods may be more appropriate than the standard inverse variance method with continuity correction.

2. Small Sample Corrections and Hedges' g

When calculating the standardized mean difference (SMD), Cohen's d has a known upward bias with small sample sizes. The bias is small but systematic: Cohen's d slightly overestimates the true SMD, especially when total sample sizes are below 20-30. Hedges' g corrects for this bias by multiplying d by a correction factor J that is slightly less than 1 for small samples and approaches 1 as sample size increases.

J = 1 - 3 / (4 * (n1 + n2) - 9) For n1 + n2 = 10: J = 0.9231 (7.7% reduction) For n1 + n2 = 20: J = 0.9605 (4.0% reduction) For n1 + n2 = 50: J = 0.9844 (1.6% reduction) For n1 + n2 = 200: J = 0.9962 (0.4% reduction)

What you should do: Always use Hedges' g (not Cohen's d) in meta-analysis. MetaReview applies the Hedges' g correction automatically when you select SMD as the effect measure. If you are extracting pre-computed SMD values from individual studies, verify whether the original authors reported Cohen's d or Hedges' g, and apply the correction if needed.

3. Unit of Analysis Errors

A unit of analysis error occurs when the same participants are counted more than once in the analysis, or when the analysis does not properly account for the study design. Common scenarios include:

Solutions: For multi-arm trials, either split the shared control group proportionally (e.g., half the control N to each comparison) or combine the treatment arms if they are similar. For cluster-randomized trials, inflate the variance by the design effect (1 + (m-1)*ICC, where m = average cluster size and ICC = intraclass correlation). For crossover trials, use the paired analysis if reported, or estimate the correlation from available data.

4. Double-Counting Participants

Beyond multi-arm trials, double-counting can occur in other ways. If the same trial is published in multiple reports (e.g., an interim analysis and a final analysis), including both would count those participants twice. Similarly, if a study reports results separately for different subgroups, including all subgroups as separate "studies" while also including the overall result would double-count every participant.

How to avoid it: During study selection, carefully track all publications from each trial (using trial registration numbers like NCT identifiers). For each unique trial, select only one publication -- typically the most complete or most recent. If different publications report different outcomes, link them to the same trial rather than treating them as independent studies.

5. Combining Different Outcome Scales Without SMD

If some studies measure depression with the Hamilton scale (0-52 range) and others with the PHQ-9 (0-27 range), you cannot pool them using mean difference. The raw mean difference in Hamilton scores is not comparable to the raw mean difference in PHQ-9 scores because the scales have different ranges, different standard deviations, and different clinically meaningful thresholds. Pooling them as if they were the same scale would produce a meaningless number.

Solution: Use the standardized mean difference (Hedges' g), which expresses each study's effect in standard deviation units. This removes the scale dependency and allows valid pooling. Alternatively, if conversion formulas between scales have been validated (e.g., HDRS to BDI conversion), you could convert all studies to a common scale and use MD -- but this introduces additional assumptions and potential error.

6. Ignoring the Direction of the Effect

When pooling studies, all effects must be coded in the same direction. If some studies report "improvement" as a positive number and others report "improvement" as a negative number (e.g., blood pressure reduction), the pooled result will be meaningless. Similarly, for binary outcomes, the comparison direction (treatment vs control) must be consistent. A calculator will happily pool mismatched directions without warning, producing a result that underestimates the true effect or even reverses its sign.

How to avoid it: Before entering data, establish a consistent convention: for example, "negative values favor treatment." Then verify each study's direction. If a study reports the effect in the opposite direction, reverse the sign (for MD/SMD) or invert the ratio (for OR/RR, swap intervention and control columns) before entering it into the calculator.

7. Using Standard Error Instead of Standard Deviation (or Vice Versa)

This is a surprisingly common data extraction error. Standard error (SE) and standard deviation (SD) are related but different: SE = SD / sqrt(n). Entering SE where SD is expected (or vice versa) will dramatically distort the effect size and its variance. A study with SD = 12 and n = 100 has SE = 1.2. Entering 1.2 where the calculator expects SD would make the study appear 10 times more precise than it actually is.

How to avoid it: During data extraction, always note whether the original paper reports SD or SE (look for the exact label, or check whether the reported value seems plausible given the sample size). If in doubt, calculate: SE should be much smaller than SD for any study with n > 4. MetaReview expects SD for continuous outcomes.

The calculator is not a safeguard against all errors. Automated tools prevent arithmetic mistakes, but conceptual errors -- wrong effect measure, double-counted participants, unit of analysis issues, mismatched directions -- require human judgment. Always have a second person verify your data extraction and analysis setup before finalizing results.

Reporting Meta-Analysis Results

Transparent, complete reporting is a cornerstone of credible meta-analysis. The PRISMA 2020 statement (Page et al., BMJ, 2021) provides the definitive checklist for reporting systematic reviews and meta-analyses. This section covers what to include in your manuscript and how to cite MetaReview when it is used as your analysis tool.

PRISMA 2020 Checklist: Key Items for Meta-Analysis

The PRISMA 2020 checklist contains 27 items. Those most relevant to the meta-analysis calculation are:

PRISMA ItemSectionWhat to Report
#13a: Synthesis methodsMethodsThe effect measure (OR, RR, MD, SMD), the statistical model (fixed-effect or random-effects), the estimator for tau-squared (DerSimonian-Laird, REML, etc.), and the software used (MetaReview).
#13b: Combining resultsMethodsHow you decided whether meta-analysis was appropriate (e.g., clinical similarity of studies), any data transformations applied (log transformation of OR/RR), and how you handled multi-arm trials or missing data.
#13c: HeterogeneityMethodsWhich heterogeneity statistics you planned to compute (I-squared, Q, tau-squared, prediction interval), the thresholds used for interpretation, and what you planned to do if heterogeneity was high (subgroup analysis, meta-regression).
#13d: Sensitivity analysisMethodsPlanned sensitivity analyses: leave-one-out, excluding high risk-of-bias studies, using different models, using different effect measures.
#13e: Publication biasMethodsMethods for assessing publication bias: funnel plot, Egger's test, trim-and-fill. Specify the minimum number of studies required (typically 10).
#21: Synthesis resultsResultsFor each meta-analysis: pooled effect estimate, 95% CI, p-value, number of studies (k), total number of participants (N), I-squared, Q with p-value, tau-squared, prediction interval. Present a forest plot for each primary outcome.
#22: Publication biasResultsFunnel plot, results of Egger's test or other bias tests, adjusted estimate from trim-and-fill if applicable.

What to Include in the Methods Section

Your methods section should allow a reader to reproduce your analysis exactly. Include the following:

  1. Effect measure and justification: "We used the odds ratio (OR) as the effect measure for binary outcomes because the included studies were a mix of case-control and cohort designs with varying baseline event rates."
  2. Statistical model: "A random-effects model using the DerSimonian-Laird estimator for between-study variance (tau-squared) was used for all analyses."
  3. Heterogeneity assessment: "Heterogeneity was assessed using I-squared, Cochran's Q test (with a significance threshold of p < 0.10), tau-squared, and the 95% prediction interval. I-squared values of 0-25%, 25-50%, 50-75%, and above 75% were interpreted as low, moderate, substantial, and considerable heterogeneity, respectively."
  4. Software: "Meta-analysis calculations were performed using MetaReview (https://metareview.cc), a free, browser-based meta-analysis tool."
  5. Sensitivity analyses: "Leave-one-out sensitivity analysis was conducted to assess the influence of individual studies on the pooled estimate. A fixed-effect model was used as a sensitivity check."
  6. Publication bias: "Publication bias was assessed visually using funnel plots and quantitatively using Egger's regression test for meta-analyses including 10 or more studies."

What to Include in the Results Section

For each meta-analysis, report the complete set of results. Here is a template paragraph that follows PRISMA 2020 conventions:

"[Number] studies involving [total participants] participants were included in the meta-analysis of [outcome]. The pooled [effect measure] was [value] (95% CI: [lower] to [upper]; p = [value]), indicating [interpretation]. Heterogeneity was [low/moderate/substantial/considerable] (I² = [value]%, Q = [value], df = [value], p = [value]; tau² = [value]). The 95% prediction interval was [lower] to [upper]. [Figure reference] presents the forest plot."

Example with real numbers:

"Eight randomized controlled trials involving 12,456 participants were included in the meta-analysis of cardiovascular events. The pooled odds ratio was 0.72 (95% CI: 0.58 to 0.89; p = 0.003), indicating that the intervention significantly reduced the odds of cardiovascular events. Heterogeneity was moderate (I² = 41%, Q = 11.9, df = 7, p = 0.10; tau² = 0.06). The 95% prediction interval was 0.38 to 1.36, suggesting that while the average effect favors the intervention, the true effect in some settings could be null or even favor the control. Figure 2 presents the forest plot."

Reporting Subgroup and Sensitivity Analyses

Subgroup results should be reported with the same level of detail as the primary analysis. Additionally, include:

For leave-one-out sensitivity analysis, report whether the pooled estimate was robust (i.e., remained significant and in the same direction when any individual study was removed) or whether certain studies had a disproportionate influence.

How to Cite MetaReview

When citing MetaReview in your manuscript, include the tool name, URL, and the date of access:

In-text: "Meta-analysis was conducted using MetaReview (https://metareview.cc; accessed [date])." Reference list: MetaReview. Free Online Meta-Analysis Tool [Internet]. Available from: https://metareview.cc [Accessed YYYY-MM-DD].

Citing the software used for meta-analysis is required by PRISMA 2020 (Item #13a) and expected by virtually all journals. It ensures reproducibility: readers and reviewers can verify your results by entering the same data into the same tool. MetaReview produces deterministic results -- the same input always produces the same output -- so any reader can independently confirm your reported pooled estimates, heterogeneity statistics, and forest plots.

Common Reporting Errors to Avoid

MetaReview helps you report correctly: The results export includes a pre-written methods paragraph and a results paragraph that follow PRISMA 2020 conventions, including all required statistics (pooled estimate, CI, p-value, I-squared, Q, tau-squared, prediction interval, model, and number of studies). Use these as starting points for your manuscript.

Calculate Your Meta-Analysis Now

MetaReview is a free online meta-analysis calculator. Compute pooled effect sizes, heterogeneity statistics, and generate forest plots in under 15 minutes. No installation, no coding, no cost.

Open MetaReview - It's Free

See a live calculation: Aspirin vs Placebo OR analysis (7 RCTs) →

Stay Updated

Get notified about new features, meta-analysis tips, and calculation guides.

No spam. Unsubscribe anytime.

Frequently Asked Questions

Is MetaReview free to use?

Yes, MetaReview is completely free to use with no restrictions. It runs entirely in your web browser -- there is no software to install, no account to create, and no usage limits. You can compute pooled effect sizes (OR, RR, MD, SMD), generate forest plots and funnel plots, run leave-one-out sensitivity analyses, assess publication bias, and export publication-ready figures at no cost. There are no hidden paywalls, no trial periods, and no feature tiers. MetaReview is free because we believe meta-analysis tools should be accessible to all researchers, including graduate students, early-career researchers, and institutions with limited budgets. The tool is supported by the research community and will remain free.

How do I calculate odds ratio in meta-analysis?

To calculate a pooled odds ratio in meta-analysis, you need the 2x2 table data from each study: the number of events and total participants in both the intervention and control groups. The odds ratio for each individual study is computed as OR = (a * d) / (b * c), where a = events in intervention, b = non-events in intervention, c = events in control, and d = non-events in control. The meta-analysis calculation works on the natural log scale: ln(OR) is computed for each study along with its variance (1/a + 1/b + 1/c + 1/d). Studies are then weighted by the inverse of their variance and combined to produce a pooled ln(OR), which is back-transformed by exponentiation to give the pooled OR with its 95% confidence interval. In MetaReview, you simply select "Odds Ratio," enter event counts and totals, and the tool performs all of these calculations automatically. The result includes the pooled OR, confidence interval, p-value, study weights, and heterogeneity statistics.

What does I-squared mean in meta-analysis?

I-squared (I²) is a heterogeneity statistic that tells you what percentage of the total variation in observed effect sizes is due to true differences between studies, as opposed to random sampling error. It ranges from 0% to 100%. An I-squared of 0% means all the variation you see is consistent with chance -- the studies are essentially measuring the same effect. An I-squared of 30% means about 30% of the observed variation reflects genuine differences across studies. The Cochrane Handbook suggests these rough benchmarks: 0-25% is low heterogeneity, 25-50% is moderate, 50-75% is substantial, and 75-100% is considerable. When I-squared is high, the pooled effect estimate may not adequately represent any single study's true effect, and you should investigate the sources of heterogeneity through subgroup analysis or meta-regression. I-squared should always be reported alongside Cochran's Q test, tau-squared, and the prediction interval for a complete picture of heterogeneity.

Can I do meta-analysis without R or Stata?

Absolutely. MetaReview is specifically designed for researchers who do not use R or Stata. It provides a visual, point-and-click interface for performing all standard meta-analysis calculations: pooled effect sizes (OR, RR, MD, SMD) using inverse variance weighting, fixed-effect and random-effects models (DerSimonian-Laird), complete heterogeneity statistics (I-squared, Cochran's Q, tau-squared, prediction intervals), forest plots, funnel plots, Egger's test for publication bias, leave-one-out sensitivity analysis, and subgroup analysis. You enter your data through a simple form or CSV import, and MetaReview handles all the statistical computation. The tool produces publication-ready forest plots that can be exported as SVG or PNG. For the vast majority of systematic reviews published in biomedical journals, MetaReview provides all the analytical capability you need. R and Stata are only necessary for advanced techniques like meta-regression, network meta-analysis, or highly custom analyses that go beyond standard pairwise meta-analysis.

How many studies do I need for a meta-analysis?

A meta-analysis can technically be computed with as few as two studies. However, the reliability and interpretability of results improve with more studies. With 2-3 studies, you can compute a pooled estimate, but the random-effects model's estimate of tau-squared will be very imprecise, and you cannot meaningfully assess heterogeneity (I-squared will have wide confidence intervals) or publication bias. With 5-10 studies, heterogeneity estimates become more stable and subgroup analysis becomes possible (though with limited power). With 10 or more studies, you can formally test for publication bias using Egger's test and interpret I-squared with reasonable confidence. The Cochrane Handbook does not specify a hard minimum; it recommends that meta-analysis should be performed when it is "sensible" -- that is, when the studies are sufficiently similar in clinical and methodological terms. Even with just 2-3 high-quality studies, formal pooling provides a more rigorous summary than narrative comparison. The key is to be transparent about the limitations when the number of studies is small.

What is the difference between fixed and random effects in meta-analysis?

The fixed-effect model assumes all studies share one common true effect size. Any variation in observed results is attributed entirely to sampling error (random chance within each study). It weights studies purely by their precision (inverse variance), giving large studies substantially more influence. The random-effects model assumes each study estimates a somewhat different true effect size, because populations, interventions, and settings differ. It adds between-study variance (tau-squared) to the weighting formula, which balances study weights: large studies still get more weight, but less disproportionately. As a result, random-effects confidence intervals are wider, reflecting the additional uncertainty from between-study variability. In practice, the random-effects model is almost always the appropriate choice because some degree of clinical heterogeneity is inevitable across independently conducted studies. When I-squared = 0%, both models produce identical results. A common and defensible approach is to use random-effects as the primary analysis and fixed-effect as a sensitivity analysis, pre-specified in your protocol.

How do I handle missing data in meta-analysis?

Missing data in meta-analysis can take several forms, each requiring a different approach. If a study does not report standard deviations but reports confidence intervals, standard errors, or p-values, you can often back-calculate the SD using standard formulas (e.g., SD = SE * sqrt(n), or deriving SE from the CI width divided by 2*1.96). If means are missing but medians are reported, estimation methods exist (e.g., Wan's method for converting median/IQR to mean/SD), though these introduce assumptions. If the required data truly cannot be recovered from the published report or supplementary materials, contact the original study authors -- response rates of 30-60% are typical. If data remains unavailable after these efforts, document the study as "excluded due to insufficient data" and assess whether its exclusion might bias the overall results (e.g., if excluded studies tend to be smaller or older). For missing participant-level data (e.g., intention-to-treat vs per-protocol populations), conduct sensitivity analyses with different assumptions: best-case (missing participants had favorable outcomes), worst-case (missing participants had unfavorable outcomes), and available-case analysis.

Can MetaReview generate a forest plot?

Yes, forest plot generation is a core feature of MetaReview. When you enter your study data and run the meta-analysis, MetaReview automatically generates a publication-ready forest plot. The plot displays each study as a row with a square (point estimate) proportional to its weight and a horizontal line (95% confidence interval). A diamond at the bottom represents the pooled effect estimate and its confidence interval. Heterogeneity statistics (I-squared, Q, tau-squared) are displayed below the plot. You can customize the forest plot's appearance: sort studies by effect size, weight, year, or entry order; toggle display of weight percentages and numerical CI values; adjust axis labels and direction indicators; and configure subgroup display with subgroup-specific diamonds. The forest plot can be exported as SVG (vector format, resolution-independent, ideal for journal submission) or PNG (raster format, suitable for presentations). For a detailed guide on creating and interpreting forest plots, see our Free Forest Plot Generator Guide.