【24h】

A hybrid approach to extract scene text from videos

机译:一种从视频中提取场景文本的混合方法

获取原文
获取原文并翻译 | 示例

摘要

With escalating claim for information indexing and retrieval a lot of attempts have been done on hauling the text from images and videos. Hauling the scene text from image and video is challenging due to complex background, changeable font size, dissimilar style, unknown layout, poor resolution and blurring, position, viewing angle and so on. The primary objective of the proposed system is to detect and haul out the text from digital videos. Through this paper we propose a hybrid approach to haul out the text from videos by integration of the two popular text extraction methods: region and connected component (CC) based method. Primarily fragment the videos into frames and acquire the key frames. Text region indicator (TRI) is being developed to figure out the text prevailing confidence and candidate region by performing binarization. Artificial Neural network (ANN) is used as the classifier to filter out the text and non-text components where Optical Character Recognition (OCR) is used for verification. Text is grouped by constructing the minimum spanning tree using the bounding box (BB) distance.
机译:随着对信息索引和检索的要求不断提高,已经进行了很多尝试来从图像和视频中提取文本。由于背景复杂,字体大小可变,样式不同,布局未知,分辨率和模糊性,位置,视角等原因,从图像和视频中拖曳场景文本具有挑战性。提出的系统的主要目标是从数字视频中检测并提取文本。通过本文,我们提出了一种混合方法,通过结合两种流行的文本提取方法:基于区域和连接分量(CC)的方法,从视频中提取文本。首先将视频分成帧并获取关键帧。正在开发文本区域指示器(TRI),以通过执行二值化来找出文本占主导地位的置信度和候选区域。人工神经网络(ANN)用作分类器,以过滤掉文本和非文本组件,其中光学字符识别(OCR)用于验证。通过使用边界框(BB)距离构造最小生成树来对文本进行分组。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号