...
首页> 外文期刊>BMC Medical Research Methodology >Reproducibility of the STARD checklist: an instrument to assess the quality of reporting of diagnostic accuracy studies
【24h】

Reproducibility of the STARD checklist: an instrument to assess the quality of reporting of diagnostic accuracy studies

机译:STARD清单的可重复性:一种评估诊断准确性研究报告质量的工具

获取原文

摘要

Background In January 2003, STAndards for the Reporting of Diagnostic accuracy studies (STARD) were published in a number of journals, to improve the quality of reporting in diagnostic accuracy studies. We designed a study to investigate the inter-assessment reproducibility, and intra- and inter-observer reproducibility of the items in the STARD statement. Methods Thirty-two diagnostic accuracy studies published in 2000 in medical journals with an impact factor of at least 4 were included. Two reviewers independently evaluated the quality of reporting of these studies using the 25 items of the STARD statement. A consensus evaluation was obtained by discussing and resolving disagreements between reviewers. Almost two years later, the same studies were evaluated by the same reviewers. For each item, percentages agreement and Cohen's kappa between first and second consensus assessments (inter-assessment) were calculated. Intraclass Correlation coefficients (ICC) were calculated to evaluate its reliability. Results The overall inter-assessment agreement for all items of the STARD statement was 85% (Cohen's kappa 0.70) and varied from 63% to 100% for individual items. The largest differences between the two assessments were found for the reporting of the rationale of the reference standard (kappa 0.37), number of included participants that underwent tests (kappa 0.28), distribution of the severity of the disease (kappa 0.23), a cross tabulation of the results of the index test by the results of the reference standard (kappa 0.33) and how indeterminate results, missing data and outliers were handled (kappa 0.25). Within and between reviewers, also large differences were observed for these items. The inter-assessment reliability of the STARD checklist was satisfactory (ICC = 0.79 [95% CI: 0.62 to 0.89]). Conclusion Although the overall reproducibility of the quality of reporting on diagnostic accuracy studies using the STARD statement was found to be good, substantial disagreements were found for specific items. These disagreements were not so much caused by differences in interpretation of the items by the reviewers but rather by difficulties in assessing the reporting of these items due to lack of clarity within the articles. Including a flow diagram in all reports on diagnostic accuracy studies would be very helpful in reducing confusion between readers and among reviewers.
机译:背景技术2003年1月,《诊断准确性研究报告标准》(STARD)在许多期刊上发表,以提高诊断准确性研究报告的质量。我们设计了一项研究,以调查STARD语句中项目的评估之间的可重复性以及观察者内部和观察者之间的可重复性。方法包括2000年在医学期刊上发表的32项诊断准确性研究,影响因子至少为4。两名审阅者使用STARD声明中的25个项目独立评估了这些研究报告的质量。通过讨论并解决审阅者之间的分歧,获得了共识评估。差不多两年后,相同的研究由相同的审阅者进行评估。对于每个项目,都计算了第一和第二次共识评估(内部评估)之间的百分比一致性和Cohen的kappa。计算类内相关系数(ICC)以评估其可靠性。结果STARD声明中所有项目的整体互评协议为85%(科恩kappa为0.70),单个项目从63%到100%不等。两项评估之间的最大差异在于参考标准的基本原理(kappa 0.37),参加测试的参与者人数(kappa 0.28),疾病严重程度分布(kappa 0.23),交叉通过参考标准的结果(kappa 0.33)对指数测试的结果进行列表,以及如何处理不确定的结果,缺失数据和异常值(kappa 0.25)。在审稿人内部和审稿人之间,这些项目也观察到很大差异。 STARD检查表的相互评估可靠性令人满意(ICC = 0.79 [95%CI:0.62至0.89])。结论尽管发现使用STARD陈述的诊断准确性研究报告的质量总体可重复性良好,但在特定项目上却存在实质性分歧。这些分歧并不是由于审稿人对项目的解释不同而引起的,而是由于条款内缺乏明确性而难以评估这些项目的报告。在所有有关诊断准确性研究的报告中包括流程图,将有助于减少读者之间和审阅者之间的混淆。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号