首页> 外文期刊>Pattern Recognition: The Journal of the Pattern Recognition Society >Improved localization accuracy by LocNet for Faster R-CNN based text detection in natural scene images
【24h】

Improved localization accuracy by LocNet for Faster R-CNN based text detection in natural scene images

机译:通过LOCNET提高本地化精度,以便在自然场景图像中更快的基于R-CNN的文本检测

获取原文
获取原文并翻译 | 示例
       

摘要

Although Faster R-CNN based text detection approaches have achieved promising results, their localization accuracy is not satisfactory in certain cases due to their sub-optimal bounding box regression based localization modules. In this paper, we address this problem and propose replacing the bounding box regression module with a novel LocNet based localization module to improve the localization accuracy of a Faster R-CNN based text detector. Given a proposal generated by a region proposal network (RPN), instead of directly predicting the bounding box coordinates of the concerned text instance, the proposal is enlarged to create a search region so that an "In-Out" conditional probability to each row and column of this search region is assigned, which can then be used to accurately infer the concerned bounding box. Furthermore, we present a simple yet effective two-stage approach to convert the difficult multi oriented text detection problem to a relatively easier horizontal text detection problem, which makes our approach able to robustly detect multi-oriented text instances with accurate bounding box localization. Experiments demonstrate that the proposed approach boosts the localization accuracy of Faster R-CNN based text detectors significantly. Consequently, our new text detector has achieved superior performance on both horizontal (ICDAR-2011, ICDAR-2013 and MULTILIGUL) and multi-oriented (MSRA-TD500, ICDAR-2015) text detection benchmark tasks. (C) 2019 Elsevier Ltd. All rights reserved.
机译:虽然基于R-CNN的文本检测方法更快地实现了有希望的结果,但它们的本地化精度在某些情况下不会令人满意,因为它们的子最优边界盒基于的本地化模块。在本文中,我们解决了这个问题,并提出了用基于​​新的基因网的定位模块更换边界框回归模块,以提高基于R-CNN的文本检测器的定位精度。给定由区域提案网络(RPN)生成的提议,而不是直接预测有关文本实例的边界框坐标,该提议扩大以创建搜索区域,以便为每行的“输入”条件概率分配了该搜索区域的列,然后可以使用该列来准确地推断有关边界。此外,我们提出了一种简单而有效的两阶段方法,将困难的多面向文本检测问题转换为相对容易的水平文本检测问题,这使得我们的方法能够通过准确的边界盒定位鲁棒地检测多面向文本实例。实验表明,所提出的方法显着提高了基于R-CNN基本探测器的定位精度。因此,我们的新文本探测器在水平(ICDAR-2011,ICDAR-2013和MultiLiligul)上实现了卓越的性能和多型(MSRA-TD500,ICDAR-2015)文本检测基准任务。 (c)2019年elestvier有限公司保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号