首页> 外文期刊>Systematic Reviews >Machine learning for screening prioritization in systematic reviews: comparative performance of Abstrackr and EPPI-Reviewer
【24h】

Machine learning for screening prioritization in systematic reviews: comparative performance of Abstrackr and EPPI-Reviewer

机译:用于筛选系统评价优先级的机器学习:Abstrackr和EPPI-Reviewer的比较表现

获取原文
           

摘要

Improving the speed of systematic review (SR) development is key to supporting evidence-based medicine. Machine learning tools which semi-automate citation screening might improve efficiency. Few studies have assessed use of screening prioritization functionality or compared two tools head to head. In this project, we compared performance of two machine-learning tools for potential use in citation screening. Using 9 evidence reports previously completed by the ECRI Institute Evidence-based Practice Center team, we compared performance of Abstrackr and EPPI-Reviewer, two off-the-shelf citations screening tools, for identifying relevant citations. Screening prioritization functionality was tested for 3 large reports and 6 small reports on a range of clinical topics. Large report topics were imaging for pancreatic cancer, indoor allergen reduction, and inguinal hernia repair. We trained Abstrackr and EPPI-Reviewer and screened all citations in 10% increments. In Task 1, we inputted whether an abstract was ordered for full-text screening; in Task 2, we inputted whether an abstract was included in the final report. For both tasks, screening continued until all studies ordered and included for the actual reports were identified. We assessed potential reductions in hypothetical screening burden (proportion of citations screened to identify all included studies) offered by each tool for all 9 reports. For the 3 large reports, both EPPI-Reviewer and Abstrackr performed well with potential reductions in screening burden of 4 to 49% (Abstrackr) and 9 to 60% (EPPI-Reviewer). Both tools had markedly poorer performance for 1 large report (inguinal hernia), possibly due to its heterogeneous key questions. Based on McNemar’s test for paired proportions in the 3 large reports, EPPI-Reviewer outperformed Abstrackr for identifying articles ordered for full-text review, but Abstrackr performed better in 2 of 3 reports for identifying articles included in the final report. For small reports, both tools provided benefits but EPPI-Reviewer generally outperformed Abstrackr in both tasks, although these results were often not statistically significant. Abstrackr and EPPI-Reviewer performed well, but prioritization accuracy varied greatly across reports. Our work suggests screening prioritization functionality is a promising modality offering efficiency gains without giving up human involvement in the screening process.
机译:提高系统评价速度(SR)开发是支持循证医学的关键。机器学习工具,半自动化引用筛选可能提高效率。很少有研究已经评估了使用筛选优先级功能或将两个工具头比较到头部。在该项目中,我们比较了两种机器学习工具的表现,以便在引用筛选中潜在使用。使用Ecri Institute循证实践中心团队完成的9个证据报告,我们比较了Abstrackr和EPPI-Reviewer,两个现成引用筛选工具的表现,以确定有关引用。在一系列临床主题的3个大型报告和6个小报告中测试了筛选优先级功能。大型报告主题正在为胰腺癌,减少室内过敏原和腹股沟疝修复成像。我们培训了Abstrackr和EPPI-Reviewer,并以10%的增量筛选了所有引文。在任务1中,我们输入了抽象是否订购了全文筛选;在任务2中,我们输入了摘要是否包含在最终报告中。对于两个任务,筛选持续,直到确定所有订购和包括实际报告的研究。我们评估了所有9个报告所提供的每个工具所提供的假设筛选负担的潜在减少(筛选的引用,以确定所有包括的研究)。对于3个大型报告,EPPI-Reviewer和Abstrackr都在筛选4到49%(Abstrackr)和9至60%(EPPI-Reviewer)的潜在减少。这两种工具对1个大型报告(Incuinal Hernia)的表现明显较差,可能是由于其异质关键问题。基于McNemar在3个大型报告中进行配对比例的测试,EPPI-Reviewer销售订购全文审查的物品的Exprackr,但Abstrackr在3个报告中更好地识别最终报告中包含的条款。对于小型报告,这两种工具都提供了福利,但EPPI-Reviewer一般在两个任务中表现出避风症,尽管这些结果往往没有统计学意义。 Abstrackr和EPPI-Reviewer表现良好,但报告中的优先级精度大大变化。我们的工作建议筛选优先级功能是一个有前途的模式,在不放弃筛选过程中的人类参与的情况下提供效率提升。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号