Risk of bias assessment of test comparisons was uncommon in comparative accuracy systematic reviews: an overview of reviews

J Clin Epidemiol. 2020 Nov:127:167-174. doi: 10.1016/j.jclinepi.2020.08.007. Epub 2020 Aug 13.

Abstract

Objectives: Comparative diagnostic test accuracy systematic reviews (DTA reviews) assess the accuracy of two or more tests and compare their diagnostic performance. We investigated how comparative DTA reviews assessed the risk of bias (RoB) in primary studies that compared multiple index tests.

Study design and setting: This is an overview of comparative DTA reviews indexed in MEDLINE from January 1st to December 31st, 2017. Two assessors independently identified DTA reviews including at least two index tests and containing at least one statement in which the accuracy of the index tests was compared. Two assessors independently extracted data on the methods used to assess RoB in studies that directly compared the accuracy of multiple index tests.

Results: We included 238 comparative DTA reviews. Only two reviews (0.8%, 95% confidence interval 0.1 to 3.0%) conducted RoB assessment of test comparisons undertaken in primary studies; neither used an RoB tool specifically designed to assess bias in test comparisons.

Conclusion: Assessment of RoB in test comparisons undertaken in primary studies was uncommon in comparative DTA reviews, possibly due to lack of existing guidance on and awareness of potential sources of bias. Based on our findings, guidance on how to assess and incorporate RoB in comparative DTA reviews is needed.

Keywords: Bias; Diagnostic accuracy; Meta-analysis; Systematic review; Test comparison.

Publication types

  • Research Support, Non-U.S. Gov't
  • Systematic Review

MeSH terms

  • Bias*
  • Confidence Intervals
  • Data Accuracy
  • Diagnostic Tests, Routine / standards*
  • Humans
  • Systematic Reviews as Topic*