Detect publication bias in your meta-analysis with contour-enhanced funnel plots showing significance regions at p<0.01, p<0.05, and p<0.10. Run Egger's regression test, Begg's rank correlation, trim-and-fill, and three fail-safe N variants (Rosenthal, Orwin with customizable trivial effect, Rosenberg). Adjust confidence levels from 80% to 99%, customize font sizes, copy the plot to clipboard, auto-generate a publication-ready methods paragraph, and export reproducible R code for the metafor package. Import from the pipeline, drag-drop CSV/Excel, or enter data manually.
Move data between tools automatically. Compute effect sizes, then send results to Forest Plot, Funnel Plot, or Heterogeneity analysis with one click.
No data in pipeline yet. Compute effect sizes or convert data in any tool, then send it downstream.
Load sample data to see how the tool works, or clear all fields to start fresh.
Trivial effect threshold for Orwin's fail-safe N
Drag & drop a file or
CSV, TSV, Excel (.xlsx/.xls) - max 500 rows
| Study | Effect Size | Standard Error | |
|---|---|---|---|
Import studies directly from the Effect Size Calculator or Forest Plot via the pipeline banner, drag-drop a CSV/Excel file, paste spreadsheet data, or add studies manually. All inputs are autosaved automatically.
The tool plots effect size (x-axis) against standard error (y-axis, inverted), with a pseudo confidence interval triangle around the pooled estimate. Use the confidence level selector to choose 80%, 90%, 95%, or 99% CI, and the font size slider to customize plot text.
Enable contour-enhanced mode to overlay significance regions at p<0.01, p<0.05, and p<0.10. See at a glance which studies fall within or outside statistical significance thresholds.
Egger's regression test, Begg's rank correlation test, trim-and-fill with imputed studies, and fail-safe N calculations (Rosenthal's, Orwin's with customizable trivial effect threshold, and Rosenberg's).
Auto-generate a publication-ready methods paragraph summarizing your bias assessment results. Copy reproducible metafor R package code for your analysis script.
Download the funnel plot as SVG or PNG, copy it to clipboard as a high-resolution PNG, or export study data as CSV or Excel. Copy test statistics for your manuscript.
Send your data forward to the Sensitivity Analysis tool, or back to the Forest Plot generator. The pipeline workflow bar shows where you are in the analysis flow.
Need this done professionally? Get a complete publication bias assessment with expert interpretation.
Get a Free QuoteFunnel plot asymmetry can result from true heterogeneity, methodological differences between small and large studies, chance, or artefacts of the effect measure. Always consider alternative explanations before concluding publication bias.
No single test is definitive. Combine visual funnel plot inspection with Egger's test, Begg's test, trim-and-fill, and fail-safe N. Sensitivity analyses (comparing fixed vs. random effects, or excluding outliers) strengthen your conclusions.
Cochrane and PRISMA guidelines require explicit reporting of publication bias assessment. Use the auto-generated methods paragraph as a starting point: it includes the tests used, test statistics, p-values, and number of imputed studies. Adjust the wording for your manuscript and discuss implications for the certainty of evidence.
When you toggle contour enhancement, significance bands at p<0.01, p<0.05, and p<0.10 reveal whether missing studies would have been non-significant (pointing to publication bias) or significant (pointing to heterogeneity or other factors). Peters et al. (2008) showed this distinction substantially improves the diagnostic accuracy of funnel plots.
The adjusted pooled estimate from trim-and-fill should be interpreted as 'what the result might look like if bias were present,' not as the true unbiased estimate. If the conclusion does not change materially, the finding is more robust.
Rosenthal's, Orwin's, and Rosenberg's fail-safe N each estimate how many unpublished null studies would need to exist to overturn your conclusion. A large fail-safe N (exceeding the 5k+10 threshold, where k is the number of studies) suggests robust findings. Orwin's variant is especially useful because you can set a custom trivial effect threshold relevant to your clinical context.
Publication bias occurs when studies with statistically significant or favorable results are more likely to be published than studies with null or negative findings. This selective reporting distorts the evidence base and inflates pooled effect estimates in meta-analysis. Beyond classic publication bias, funnel plot asymmetry can also arise from time-lag bias (where studies with positive results are published faster) and language bias (where English-language journals preferentially publish significant findings). The Cochrane Handbook for Systematic Reviews of Interventions (Higgins et al., 2023, Chapter 13) identifies publication bias as one of the most serious threats to the validity of systematic reviews, and PRISMA 2020 reporting guidelines require explicit assessment.
A funnel plot generator creates the primary visual diagnostic for detecting this bias. The funnel plot maps each study's effect size on the horizontal axis against its standard error on the vertical axis (with the most precise studies at the top). In the absence of bias, studies scatter symmetrically around the pooled estimate, forming an inverted funnel shape. The confidence level selector lets you choose between 80%, 90%, 95%, and 99% confidence intervals for the pseudo-CI triangle, matching the precision envelope to your analysis requirements. A font size slider allows you to customize all plot text for presentations or manuscript figures. Studies falling outside the confidence region may reflect genuine heterogeneity or reporting bias. P-curve analysis provides a complementary approach by examining the distribution of statistically significant p-values to detect whether the literature contains evidential value or shows signs of p-hacking.
Standard funnel plots show whether asymmetry exists but cannot distinguish why it exists. Contour-enhanced funnel plots (Peters et al., 2008) solve this problem by overlaying shaded bands representing regions of statistical significance, typically at p<0.01, p<0.05, and p<0.10. Each band corresponds to a two-sided z-test threshold: studies falling inside the most central band are non-significant at p>0.10, while those in progressively outer bands reach conventional levels of significance. This visual stratification transforms the diagnostic question from "is there asymmetry?" to "where are the missing studies?" If the gap in the funnel falls predominantly in the non-significant region (p>0.05), publication bias is the most parsimonious explanation: reviewers selectively failed to publish null findings. Conversely, if the missing studies would have fallen in zones of statistical significance, the asymmetry is more likely driven by between-study heterogeneity, differences in study design, or artefacts of the chosen effect measure rather than selective publication.
This tool lets you toggle contour enhancement on and off to compare the standard and contour-enhanced views side by side. The significance regions are computed from the pooled effect estimate and each study's standard error, using the conventional normal approximation. We recommend interpreting the contour-enhanced plot alongside Egger's test and trim-and-fill: if trim-and-fill imputed studies land in the non-significant band, this corroborates a publication-bias explanation. If they land in significant bands, reconsider whether true heterogeneity or methodological quality differences between small and large studies are the more likely driver.
The fail-safe N (also called the file-drawer number) estimates how many unpublished studies with null results would be needed to reduce the pooled effect to non-significance or a trivial level, directly quantifying the robustness of your meta-analysis to the file-drawer problem. This tool computes three complementary variants. Rosenthal's fail-safe N (Rosenthal, 1979) is the classic approach based on summed z-values and provides a straightforward count. Orwin's fail-safe N (Orwin, 1983) is more flexible because you can set a custom trivial effect threshold that is clinically meaningful for your research question, for example, a standardized mean difference of 0.10. Rosenberg's fail-safe N (Rosenberg, 2005) is a weighted variant that accounts for sample size differences across studies, producing a more conservative estimate. A large fail-safe N relative to the number of included studies (commonly judged by the 5k+10 rule, where k is the number of studies) suggests the pooled finding is robust.
Publication bias assessment rarely stands alone; it is part of a broader meta-analysis workflow. This tool integrates with our analysis pipeline: the pipeline workflow bar at the top of the tool shows where funnel plot generation sits relative to the other stages, typically after effect size calculation and forest plot visualization, and before or alongside sensitivity analysis. You can import studies directly from the Effect Size Calculator or Forest Plot Generator via the pipeline banner, eliminating the need to re-enter data. Alternatively, drag-drop a CSV or Excel file, or paste tab-separated data copied from a spreadsheet. All inputs are autosaved to your browser, so you can close the tab and return later without losing work. When you are done, export the plot as PNG or SVG, copy the plot to clipboard as a high-resolution PNG with one click, or export the underlying study data and test statistics as CSV or Excel for use in R, Stata, or your manuscript supplementary materials.
Writing up publication bias assessment for a manuscript can be time-consuming. The auto-generated methods paragraph produces publication-ready text that summarizes the funnel plot assessment, lists the statistical tests performed, and reports their results (test statistics, p-values, number of imputed studies from trim-and-fill, and fail-safe N values). The paragraph follows standard reporting conventions recommended by PRISMA 2020 and the Cochrane Handbook, giving you a solid starting point that you can refine for your specific manuscript. For full reproducibility, the R code generator produces a ready-to-run script for the metafor package (Viechtbauer, 2010), the most widely cited meta-analysis package in R. The generated code includes your study data, funnel plot rendering, Egger's regression test, Begg's rank correlation, and trim-and-fill analysis. Copy the script with one click and paste it directly into RStudio to reproduce or extend your analysis, satisfying reviewer and journal requirements for transparent analysis scripts.
Visual inspection alone is subjective, which is why a publication bias test online tool should include formal statistical tests. Egger's regression test (Egger et al., 1997) regresses standardized effects against precision and tests whether the intercept differs from zero. A significant result (p < 0.10, using the conventional threshold for this test) indicates funnel plot asymmetry. The Egger's test calculator in this tool reports the intercept, t-statistic, and exact p-value. Begg and Mazumdar's rank correlation test provides a complementary non-parametric assessment using Kendall's tau, testing whether effect sizes are correlated with their variances.
When asymmetry is detected, the trim and fill calculator implements Duval and Tweedie's (2000) non-parametric method to estimate the number of missing studies and impute them as mirror images of the most extreme observed studies. Modern regression-based alternatives include PET-PEESE (Stanley and Doucouliagos, 2014), which uses precision-effect testing and precision-effect estimate with standard error to provide bias-corrected pooled estimates, and selection models (Vevea and Hedges, 1995), which explicitly model the probability of publication as a function of statistical significance. The adjusted pooled estimate provides a sensitivity analysis: if the conclusion changes materially after imputation, publication bias is a concern that should be discussed when rating the certainty of evidence using the GRADE framework.
Important methodological caveats apply: all tests for funnel plot asymmetry have limited statistical power with fewer than 10 studies. Funnel plot asymmetry can arise from sources other than publication bias, including genuine heterogeneity, poor methodological quality of smaller studies, or artefacts of the chosen effect measure (Sterne et al., 2011). Always interpret bias tests alongside visual inspection and clinical context. Before running bias assessment, ensure your individual study effect sizes are correctly computed using our effect size calculator, and visualize the full meta-analysis with our forest plot generator to assess the overall pattern of results. Test the stability of your pooled estimate with our leave-one-out sensitivity analysis.
A funnel plot is a scatter plot used to detect publication bias in meta-analysis. It plots each study's effect size on the x-axis against a measure of precision (typically standard error) on the y-axis, with the most precise studies at the top. In the absence of bias, studies scatter symmetrically around the pooled estimate, forming an inverted funnel shape. Asymmetry suggests that smaller studies with non-significant results may be missing, a hallmark of publication bias.
Egger's test is a statistical test for funnel plot asymmetry. It regresses the standardized effect (effect / SE) against precision (1/SE). If the intercept is significantly different from zero (p < 0.10), this indicates asymmetry consistent with publication bias. The test has adequate power when there are at least 10 studies in the meta-analysis. For fewer studies, visual inspection and other methods should be used alongside Egger's test.
Trim-and-fill is a non-parametric method that estimates the number of missing studies due to publication bias and imputes them. The algorithm iteratively: (1) trims the most extreme small studies from the asymmetric side, (2) re-estimates the pooled effect, and (3) fills in the presumed missing studies as mirror images of the trimmed studies. The adjusted pooled estimate provides a sensitivity analysis. If the conclusion changes substantially, publication bias is a concern.
Begg and Mazumdar's rank correlation test examines the association between effect sizes and their variances using Kendall's tau. A significant correlation (p < 0.10) suggests that larger effects are associated with smaller (less precise) studies, which is a pattern consistent with publication bias. Like Egger's test, Begg's test has limited power with fewer than 10 studies and should be interpreted alongside visual assessment.
Fail-safe N estimates how many unpublished studies with null results would be needed to reduce the pooled effect to non-significance or a trivial level. This tool calculates three variants: Rosenthal's fail-safe N (the classic approach based on summed z-values), Orwin's fail-safe N (which lets you set a custom trivial effect threshold, making it more flexible for clinical interpretation), and Rosenberg's fail-safe N (a weighted variant that accounts for sample size differences across studies). A large fail-safe N relative to the number of included studies (commonly judged by the 5k+10 rule, where k is the number of studies) suggests the pooled finding is robust to the file-drawer problem.
Publication bias tests have limited statistical power with fewer than 10 studies. Cochrane Handbook recommends not using funnel plots or statistical tests for asymmetry when there are fewer than 10 studies. With 10 to 30 studies, both visual inspection and statistical tests are informative. Visual funnel plot assessment remains useful even when statistical tests are underpowered, as clear asymmetry patterns can still be detected. Fail-safe N calculations can be computed with any number of studies, but their interpretation is most meaningful when paired with other bias assessments.
A minimum of 10 studies is generally recommended for funnel plot visual inspection and formal tests like Egger's regression. With fewer than 10 studies, the power to detect asymmetry is too low, and both visual assessment and statistical tests become unreliable. The Cochrane Handbook advises against using funnel plots with fewer than 10 studies.
Funnel plot asymmetry can result from several non-publication-bias sources: true heterogeneity (smaller studies conducted in different populations), methodological differences correlated with study size, chance variation with few studies, language bias, or selective outcome reporting within studies. Sterne et al. (2011) emphasized that asymmetry should not be automatically equated with publication bias.
A contour-enhanced funnel plot (Peters et al., 2008) overlays shaded regions of statistical significance onto the standard funnel plot. The regions typically correspond to p<0.01, p<0.05, and p<0.10. If the 'missing' studies fall predominantly in the non-significant region, publication bias is the likely explanation for asymmetry. If the missing studies would fall in zones of statistical significance, the asymmetry is more likely due to between-study heterogeneity or other factors rather than selective publication of significant results. This tool lets you toggle contour enhancement on and off, and you can adjust the confidence level (80%, 90%, 95%, or 99%) to match the precision envelope to your analysis requirements.
Yes. The R code generator produces ready-to-run code for the metafor package (Viechtbauer, 2010), the most widely used meta-analysis package in R. The generated script includes your study data, funnel plot rendering, Egger's regression test, Begg's rank correlation test, and trim-and-fill analysis. You can copy the code with one click and paste it directly into RStudio or any R environment. This is useful for reproducibility, for running additional analyses beyond what the browser tool offers, or for satisfying reviewer requests to provide analysis scripts.
After running the publication bias tests, the tool generates a publication-ready methods paragraph that summarizes the funnel plot assessment, statistical tests performed, and their results. The paragraph follows standard reporting conventions recommended by PRISMA 2020 and the Cochrane Handbook, including the specific tests used, their test statistics, p-values, and the number of imputed studies from trim-and-fill. You can copy this paragraph directly into your manuscript's methods or results section and adjust the wording as needed.
Trim-and-fill (Duval and Tweedie, 2000) is a non-parametric method that estimates the number of missing studies from funnel plot asymmetry, imputes their effect sizes, and recalculates the pooled estimate. It is a sensitivity analysis tool, not a definitive correction. The adjusted estimate shows what the result might look like if bias were removed, but it assumes the asymmetry is entirely due to publication bias.
Yes. The confidence level selector lets you choose between 80%, 90%, 95%, and 99% confidence intervals for the pseudo-CI triangle displayed on the funnel plot. The default is 95%, which matches the most common convention in published systematic reviews. Adjusting the confidence level changes the width of the expected funnel boundary, which can be useful when exploring whether studies fall within or outside tighter or wider precision envelopes.
Visualize individual study estimates with our forest plot generator for meta-analysis with weighted squares, CI lines, and diamond summary. Test the robustness of your pooled estimate with our leave-one-out sensitivity analysis simulator. Calculate individual study effect sizes before plotting with our effect size calculator for SMD, OR, and RR. For small meta-analyses where funnel plots have limited power, our Doi plot with LFK index offers a more sensitive measure of asymmetry. Assess whether the significant findings in your review reflect genuine effects or selective reporting with our p-curve analysis tool.
Reviewed by
Dr. Sarah Mitchell holds a PhD in Biostatistics from Johns Hopkins Bloomberg School of Public Health and has over 15 years of experience in systematic review methodology and meta-analysis. She has authored or co-authored 40+ peer-reviewed publications in journals including the Journal of Clinical Epidemiology, BMC Medical Research Methodology, and Research Synthesis Methods. A former Cochrane Review Group statistician and current editorial board member of Systematic Reviews, Dr. Mitchell has supervised 200+ evidence synthesis projects across clinical medicine, public health, and social sciences. She reviews all Research Gold tools to ensure statistical accuracy and compliance with Cochrane Handbook and PRISMA 2020 standards.
Our PhD team runs complete meta-analyses: data extraction, effect size computation, forest plots, sensitivity analysis, and a manuscript ready for journal submission. Average turnaround: 2-4 weeks.