【24h】

Cross-modal Memory Networks for Radiology Report Generation

机译:放射学报告生成的跨模型内存网络

获取原文

摘要

Medical imaging plays a significant role in clinical practice of medical diagnosis, where the text reports of the images are essential in understanding them and facilitating later treatments. By generating the reports automatically, it is beneficial to help lighten the burden of radiologists and significantly promote clinical automation, which already attracts much attention in applying artificial intelligence to medical domain. Previous studies mainly follow the encoder-decoder paradigm and focus on the aspect of text generation, with few studies considering the importance of cross-modal mappings and explicitly exploit such mappings to facilitate radiology report generation. In this paper, we propose a cross-modal memory networks (CMN) to enhance the encoder-decoder framework for radiology report generation, where a shared memory is designed to record the alignment between images and texts so as to facilitate the interaction and generation across modalities. Experimental results illustrate the effectiveness of our proposed model, where state-of-the-art performance is achieved on two widely used benchmark datasets. i.e., IU X-Ray and MIMIC-CXR. Further analyses also prove that our model is able to better align information from radiology images and texts so as to help generating more accurate reports in terms of clinical indicators.
机译:医学成像在医学诊断的临床实践中起着重要作用,其中图像的文本报告对于了解它们并促进后期治疗至关重要。通过自动生成报告,有助于减轻放射科医师的负担,并显着促进临床自动化,这已经吸引了对向医疗领域应用人工智能的关注。以前的研究主要遵循编码器 - 解码器范例并专注于文本生成的方面,几乎没有考虑跨模型映射的重要性,并明确利用这种映射以促进放射学报告生成。在本文中,我们提出了一种跨模型存储器网络(CMN)来增强用于放射学报告生成的编码器 - 解码器框架,其中共享存储器被设计为记录图像和文本之间的对准,以便于跨处理和产生方式。实验结果说明了我们所提出的模型的有效性,其中在两个广泛使用的基准数据集中实现了最先进的性能。即,IU X射线和模拟-CXR。进一步的分析还证明我们的模型能够更好地从放射学图像和文本对准信息,以帮助在临床指标方面产生更准确的报告。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号