...
首页> 外文期刊>Circuits and Systems for Video Technology, IEEE Transactions on >Multioriented Video Scene Text Detection Through Bayesian Classification and Boundary Growing
【24h】

Multioriented Video Scene Text Detection Through Bayesian Classification and Boundary Growing

机译:贝叶斯分类和边界增长的多方位视频场景文本检测

获取原文
获取原文并翻译 | 示例

摘要

Multioriented text detection in video frames is not as easy as detection of captions or graphics or overlaid texts, which usually appears in the horizontal direction and has high contrast compared to its background. Multioriented text generally refers to scene text that makes text detection more challenging and interesting due to unfavorable characteristics of scene text. Therefore, conventional text detection methods may not give good results for multioriented scene text detection. Hence, in this paper, we present a new enhancement method that includes the product of Laplacian and Sobel operations to enhance text pixels in videos. To classify true text pixels, we propose a Bayesian classifier without assuming a priori probability about the input frame but estimating it based on three probable matrices. Three different ways of clustering are performed on the output of the enhancement method to obtain the three probable matrices. Text candidates are obtained by intersecting the output of the Bayesian classifier with the Canny edge map of the input frame. A boundary growing method is introduced to traverse the multioriented scene text lines using text candidates. The boundary growing method works based on the concept of nearest neighbors. The robustness of the method has been tested on a variety of datasets that include our own created data (nonhorizontal and horizontal text data) and two publicly available data, namely, video frames of Hua and complex scene text data of ICDAR 2003 competition (camera images). Experimental results show that the performance of the proposed method is encouraging compared with results of existing methods in terms of recall, precision, F-measures, and computational times.
机译:视频帧中的多方向文本检测不如字幕或图形或重叠文本检测那么容易,字幕或图形或覆盖文本通常在水平方向上显示,并且与背景相比具有很高的对比度。多向文本通常是指场景文本,由于场景文本的不利特性,使得文本检测更具挑战性和趣味性。因此,传统的文本检测方法可能无法为多方位场景文本检测提供良好的结果。因此,在本文中,我们提出了一种新的增强方法,其中包括拉普拉斯算子和Sobel运算的乘积,以增强视频中的文本像素。为了对真实文本像素进行分类,我们提出了一种贝叶斯分类器,而无需假设输入帧具有先验概率,而是基于三个可能的矩阵对其进行估计。对增强方法的输出执行三种不同的聚类方法,以获得三个可能的矩阵。通过将贝叶斯分类器的输出与输入帧的Canny边缘图相交而获得文本候选。引入边界增长方法,以使用文本候选遍历多方位场景文本行。边界增长方法基于最近邻的概念。该方法的鲁棒性已经在包括我们自己创建的数据(非水平和水平文本数据)以及两个公开可用的数据(包括Hua的视频帧和ICDAR 2003竞赛的复杂场景文本数据(相机图像))的各种数据集上进行了测试。 )。实验结果表明,在召回率,精度,F度量和计算时间方面,与现有方法相比,该方法的性能令人鼓舞。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号