首页> 外文会议>European conference on computer vision >Detecting Text in Natural Image with Connectionist Text Proposal Network
【24h】

Detecting Text in Natural Image with Connectionist Text Proposal Network

机译:使用Connectionist文本提议网络检测自然图像中的文本

获取原文

摘要

We propose a novel Connectionist Text Proposal Network (CTPN) that accurately localizes text lines in natural image. The CTPN detects a text line in a sequence of fine-scale text proposals directly in convolutional feature maps. We develop a vertical anchor mechanism that jointly predicts location and texton-text score of each fixed-width proposal, considerably improving localization accuracy. The sequential proposals are naturally connected by a recurrent neural network, which is seamlessly incorporated into the convolutional network, resulting in an end-to-end trainable model. This allows the CTPN to explore rich context information of image, making it powerful to detect extremely ambiguous text. The CTPN works reliably on multi-scale and multi-language text without further post-processing, departing from previous bottom-up methods requiring multi-step post filtering. It achieves 0.88 and 0.61 F-measure on the ICDAR 2013 and 2015 benchmarks, surpassing recent results by a large margin. The CTPN is computationally efficient with 0.14 s/image, by using the very deep VGG16 model.
机译:我们提出了一种新颖的连接主义者文本提案网络(CTPN),该网络可以准确地定位自然图像中的文本行。 CTPN直接在卷积特征图中检测一系列精细文本建议中的文本行。我们开发了一种垂直定位机制,可以共同预测每个固定宽度提案的位置和文本/非文本得分,从而大大提高定位精度。顺序提议通过递归神经网络自然连接,该递归神经网络被无缝地集成到卷积网络中,从而形成了端到端的可训练模型。这使CTPN能够探索图像的丰富上下文信息,从而使其能够强大地检测出极端模棱两可的文本。 CTPN可以可靠地在多尺度和多语言文本上运行,而无需进一步的后处理,这与以前的自下而上的方法(需要多步后过滤)不同。在ICDAR 2013和2015基准上,它达到了0.88和0.61 F值,大大超过了最近的结果。通过使用非常深的VGG16模型,CTPN的计算效率为0.14 s /图像。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号