首页> 外文会议>International Conference on Medical Image Computing and Computer-Assisted Intervention >Automatic Radiology Report Generation Based on Multi-view Image Fusion and Medical Concept Enrichment
【24h】

Automatic Radiology Report Generation Based on Multi-view Image Fusion and Medical Concept Enrichment

机译:基于多视图图像融合和医学概念充实的放射线报告自动生成

获取原文

摘要

Generating radiology reports is time-consuming and requires extensive expertise in practice. Therefore, reliable automatic radiology report generation is highly desired to alleviate the workload. Although deep learning techniques have been successfully applied to image classification and image captioning tasks, radiology report generation remains challenging in regards to understanding and linking complicated medical visual contents with accurate natural language descriptions. In addition, the data scales of open-access datasets that contain paired medical images and reports remain very limited. To cope with these practical challenges, we propose a generative encoder-decoder model and focus on chest x-ray images and reports with the following improvements. First, we pretrain the encoder with a large number of chest x-ray images to accurately recognize 14 common radiographic observations, while taking advantage of the multi-view images by enforcing the cross-view consistency. Second, we synthesize multi-view visual features based on a sentence-level attention mechanism in a late fusion fashion. In addition, in order to enrich the decoder with descriptive semantics and enforce the correctness of the deterministic medical-related contents such as mentions of organs or diagnoses, we extract medical concepts based on the radiology reports in the training data and fine-tune the encoder to extract the most frequent medical concepts from the x-ray images. Such concepts are fused with each decoding step by a word-level attention model. The experimental results conducted on the Indiana University Chest X-Ray dataset demonstrate that the proposed model achieves the state-of-the-art performance compared with other baseline approaches.
机译:生成放射学报告非常耗时,并且在实践中需要广泛的专业知识。因此,非常需要可靠的自动放射学报告生成以减轻工作量。尽管深度学习技术已成功地应用于图像分类和图像字幕任务,但放射学报告的生成在理解和链接具有精确自然语言描述的复杂医学视觉内容方面仍然具有挑战性。此外,包含成对的医学图像和报告的开放获取数据集的数据规模仍然非常有限。为了应对这些实际挑战,我们提出了一种生成式编码器/解码器模型,并着重于胸部X射线图像和报告,并进行了以下改进。首先,我们使用大量的胸部X射线图像对编码器进行预训练,以准确识别14种常见的射线照相观察结果,同时通过增强交叉视图的一致性来利用多视图图像。其次,我们以后期融合的方式基于句子级别的注意力机制来合成多视图视觉特征。另外,为了使解码器具有描述性语义,并增强确定性医学相关内容(例如提及器官或诊断)的正确性,我们根据训练数据中的放射学报告提取医学概念,并对编码器进行微调从X射线图像中提取最常见的医学概念。这些概念通过单词级别的关注模型与每个解码步骤融合在一起。在印第安纳大学胸部X射线数据集上进行的实验结果表明,与其他基准方法相比,该模型可实现最先进的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号