首页> 外文期刊>BMC Bioinformatics >The gene normalization task in BioCreative III
【24h】

The gene normalization task in BioCreative III

机译:生物重建III中的基因标准化任务

获取原文
获取外文期刊封面目录资料

摘要

BackgroundWe report the Gene Normalization (GN) challenge in BioCreative III where participating teams were asked to return a ranked list of identifiers of the genes detected in full-text articles. For training, 32 fully and 500 partially annotated articles were prepared. A total of 507 articles were selected as the test set. Due to the high annotation cost, it was not feasible to obtain gold-standard human annotations for all test articles. Instead, we developed an Expectation Maximization (EM) algorithm approach for choosing a small number of test articles for manual annotation that were most capable of differentiating team performance. Moreover, the same algorithm was subsequently used for inferring ground truth based solely on team submissions. We report team performance on both gold standard and inferred ground truth using a newly proposed metric called Threshold Average Precision (TAP-k).ResultsWe received a total of 37 runs from 14 different teams for the task. When evaluated using the gold-standard annotations of the 50 articles, the highest TAP-k scores were 0.3297 (k=5), 0.3538 (k=10), and 0.3535 (k=20), respectively. Higher TAP-k scores of 0.4916 (k=5, 10, 20) were observed when evaluated using the inferred ground truth over the full test set. When combining team results using machine learning, the best composite system achieved TAP-k scores of 0.3707 (k=5), 0.4311 (k=10), and 0.4477 (k=20) on the gold standard, representing improvements of 12.4%, 21.8%, and 26.6% over the best team results, respectively.ConclusionsBy using full text and being species non-specific, the GN task in BioCreative III has moved closer to a real literature curation task than similar tasks in the past and presents additional challenges for the text mining community, as revealed in the overall team results. By evaluating teams using the gold standard, we show that the EM algorithm allows team submissions to be differentiated while keeping the manual annotation effort feasible. Using the inferred ground truth we show measures of comparative performance between teams. Finally, by comparing team rankings on gold standard vs. inferred ground truth, we further demonstrate that the inferred ground truth is as effective as the gold standard for detecting good team performance.
机译:Backgroundwe报告了生物重建III中的基因标准化(GN)挑战,其中被要求参与团队返回全文文章中检测到的基因的标识符列表。对于培训,准备32个完整和500个部分注释的物品。共选择507篇文章作为测试集。由于高的注释成本,获得所有测试物品的金标人注释是不可行的。相反,我们开发了一个期望最大化(EM)算法方法,用于选择少量测试文章的用于手动注释,这些方法最能够区分团队性能。此外,随后用于仅基于团队提交的地面真理来使用相同的算法。我们在黄金标准和使用名为阈值平均精度(TAP-K)的新提出的指标推断出基本真理的团队表现.ResultWe总共收到的37个来自14个不同团队的任务。当使用50制品的金标标注释评估时,最高的TAP-K分数分别为0.3297(k = 5),0.3538(k = 10)和0.3535(k = 20)。在使用完整的测试集中评估时,观察到较高的TAP-K分数为0.4916(k = 5,10,20)。当使用机器学习结合团队结果时,最佳复合系统在黄金标准上实现了0.3707(k = 5),0.4311(k = 10)和0.4477(k = 20)的TAP-K分数,代表12.4%的改善,在最佳团队结果中,21.8%和26.6%。使用全文和物种非特定的组合,生物重建III的GN任务比过去类似的任务更接近真正的文献策划任务,并提出了额外的挑战对于文本挖掘社区,正如整个团队的结果所透露。通过使用黄金标准进行评估团队,我们表明EM算法允许在保持手动注释工作的同时对团队提交进行区分。使用推断的地面真理,我们展示了团队之间的比较绩效措施。最后,通过比较黄金标准与推断的基础真理的团队排名,我们进一步证明了推断的实际真理与检测良好团队表现的黄金标准有效。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号