首页> 外文期刊>Journal of the Association for Information Science and Technology >Unified deep neural network for segmentation and labeling of multipanel biomedical figures
【24h】

Unified deep neural network for segmentation and labeling of multipanel biomedical figures

机译:统一的深神经网络,用于分割和标记多工具生物医学人物

获取原文
获取原文并翻译 | 示例
           

摘要

Recent efforts in biomedical visual question answering (VQA) research rely on combined information gathered from the image content and surrounding text supporting the figure. Biomedical journals are a rich source of information for such multimodal content indexing. For multipanel figures in these journals, it is critical to develop automatic figure panel splitting and label recognition algorithms to associate individual panels with text metadata in the figure caption and the body of the article. Challenges in this task include large variations in figure panel layout, label location, size, contrast to background, and so on. In this work, we propose a deep convolutional neural network, which splits the panels and recognizes the panel labels in a single step. Visual features are extracted from several layers at various depths of the backbone neural network and organized to form a feature pyramid. These features are fed into classification and regression networks to generate candidates of panels and their labels. These candidates are merged to create the final panel segmentation result through a beam search algorithm. We evaluated the proposed algorithm on the ImageCLEF data set and achieved better performance than the results reported in the literature. In order to thoroughly investigate the proposed algorithm, we also collected and annotated our own data set of 10,642 figures. The experiments, trained on 9,642 figures and evaluated on the remaining 1,000 figures, show that combining panel splitting and panel label recognition mutually benefit each other.
机译:生物医学视觉问题回答(VQA)研究的最新努力取决于从图像内容和支持该图的周围文本收集的组合信息。生物医学期刊是此类多模式内容索引的丰富信息来源。对于这些期刊中的多工资图,至关重要的是,开发自动图形面板拆分和标签识别算法,将单个面板与图形标题和文章正文中的文本元数据相关联。此任务中的挑战包括图面板布局,标签位置,尺寸,与背景形成对比等的大变化。在这项工作中,我们提出了一个深度卷积神经网络,该网络将面板拆分并识别面板标签。视觉特征是从主干神经网络各个深度的几层中提取的,并组织形成特征金字塔。这些功能被馈入分类和回归网络,以生成面板及其标签的候选。合并这些候选物来通过梁搜索算法创建最终的面板分割结果。我们评估了ImageClef数据集的提出算法,并比文献中报道的结果更好。为了彻底研究所提出的算法,我们还收集并注释了自己的10,642个数据集。实验对9,642个数字进行了训练,并对其余1,000个数字进行了评估,表明将面板拆分和面板标签识别相互互惠互利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号