首页> 外文会议>IAPR International Conference on Document Analysis and Recognition >A Unified Video Text Detection Method with Network Flow
【24h】

A Unified Video Text Detection Method with Network Flow

机译:具有网络流的统一视频文本检测方法

获取原文

摘要

Scene text detection in videos has many application needs but has drawn less attention than that in images. Existing methods for video text detection perform unsatisfactorily because of the insufficient utilization of spatial and temporal information. In this paper, we propose a novel video text detection method with network flow based tracking. The system first applies a newly proposed Fully Convolutional Neural Network (FCN) based scene text detection method to detect texts in individual frames and then track proposals in adjacent frames with a motion-based method. Next, the text association problem is formulated into a cost-flow network and text trajectories are derived from the network with a min-cost flow algorithm. At last, the trajectories are post-processed to improve the precision accuracy. The method can detect multi-oriented scene text in videos and incorporate spatial and temporal information efficiently. Experimental results show that the method improves the detection performance remarkably on benchmark datasets, e.g., by a 15.66% increase of ATA Average Tracking Accuracy) on ICDAR video scene text dataset.
机译:视频中的场景文本检测具有许多应用需求,但引起的关注比图像中的要少。由于对空间和时间信息的利用不足,用于视频文本检测的现有方法不能令人满意地执行。在本文中,我们提出了一种基于网络流跟踪的新型视频文本检测方法。该系统首先应用一种新提出的基于全卷积神经网络(FCN)的场景文本检测方法来检测单个帧中的文本,然后使用基于运动的方法跟踪相邻帧中的提议。接下来,将文本关联问题表述为成本流网络,并使用最小成本流算法从网络中得出文本轨迹。最后对轨迹进行后处理,以提高精度。该方法可以检测视频中的多方位场景文本并且有效地合并空间和时间信息。实验结果表明,该方法显着提高了基准数据集的检测性能,例如,对ICDAR视频场景文本数据集的ATA平均跟踪精度提高了15.66%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号