首页> 外文会议>International Conference on Digital Image Computing: Techniques and Applications >Training Region Selector for Gram Stained Slides with Limited Data: A Data Distillation Approach
【24h】

Training Region Selector for Gram Stained Slides with Limited Data: A Data Distillation Approach

机译:训练区域选择器为克染色幻灯片,具有有限的数据:数据蒸馏方法

获取原文

摘要

In this paper, we tackle the region selection task for the purpose of whole slide image (WSI) analysis with a focus on Gram stained slides. Typically, the number of tiles required to capture using high magnification objective is extremely large. Thus, in practice, we need to perform region of interest selection in low magnification objectives to choose only the best candidate regions for the later high magnification analysis. With the fast development of computer vision and deep learning, it is possible to train an accurate convolutional neural network (CNN) based classifier to do this region selection task which would normally be done by experienced scientists and pathologists. However, data collection and labelling during the training stage are very time-consuming, labour intensive and expensive tasks. Researchers starting to ask a much more challenging question: how can we use unlabeled data to boost the system performance? In this paper, to answer this question and to reduce labelling effort, we propose a data distillation training framework to train a CNN classifier on limited labeled data with the help of unlabeled data. The unlabeled data will be carefully selected by a teacher model via a data distillation process and put into the training set for a student model. Extensive experiments show that the proposed framework achieves a notable gain in accuracy.
机译:在本文中,我们在整个幻灯片(WSI)分析的目的中解决了区域选择任务,重点在克染色的幻灯片上。通常,使用高放大镜所需的瓷砖的数量非常大。因此,在实践中,我们需要在低放大镜中执行感兴趣的区域区域,以仅选择稍后的高放大率分析的最佳候选区域。随着计算机愿景和深度学习的快速发展,可以培训基于准确的卷积神经网络(CNN)的分类器,以进行这项区域选择任务,通常由经验丰富的科学家和病理学家完成。但是,在培训阶段期间的数据收集和标签非常耗时,劳动密集型和昂贵的任务。研究人员开始提出更具挑战性的问题:我们如何使用未标记的数据来提高系统性能?在本文中,为了回答这个问题并减少标签努力,我们提出了一种数据蒸馏训练框架,用于在未标记的数据的帮助下培训CNN分类器上有限标记的数据。教师模型通过数据蒸馏过程仔细选择未标记的数据,并将其放入学生模型的培训集中。广泛的实验表明,所提出的框架精确地实现了显着的增益。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号