首页> 外文会议>International Conference on Document Analysis and Recognition >Feature Representations for Scene Text Character Recognition: A Comparative Study
【24h】

Feature Representations for Scene Text Character Recognition: A Comparative Study

机译:场景文本字符识别的特征表示:比较研究

获取原文

摘要

Recognizing text character from natural scene images is a challenging problem due to background interferences and multiple character patterns. Scene Text Character (STC) recognition, which generally includes feature representation to model character structure and multi-class classification to predict label and score of character class, mostly plays a significant role in word-level text recognition. The contribution of this paper is a complete performance evaluation of image-based STC recognition, by comparing different sampling methods, feature descriptors, dictionary sizes, coding and pooling schemes, and SVM kernels. We systematically analyze the impact of each option in the feature representation and classification. The evaluation results on two datasets CHARS74K and ICDAR2003 demonstrate that Histogram of Oriented Gradient (HOG) descriptor, soft-assignment coding, max pooling, and Chi-Square Support Vector Machines (SVM) obtain the best performance among local sampling based feature representations. To improve STC recognition, we apply global sampling feature representation. We generate Global HOG (GHOG) by computing HOG descriptor from global sampling. GHOG enables better character structure modeling and obtains better performance than local sampling based feature representations. The GHOG also outperforms existing methods in the two benchmark datasets.
机译:由于背景干扰和多个字符模式,识别自然场景图像的文本性质是一个具有挑战性的问题。场景文本字符(STC)识别通常包括模型字符结构和多级分类的特征表示,以预测字符类的标签和分数,大多数在字级文本识别中发挥着重要作用。本文的贡献是通过比较不同的采样方法,特征描述符,字典大小,编码,编码和池方案和SVM内核来对基于图像的STC识别的完整性能评估。我们系统地分析了每个选项在特征表示和分类中的影响。两个数据集CHARS74K和ICDAR2003上的评估结果表明,面向梯度(HOG)描述符,软分配编码,MAX池和CHI-Square支持向量机(SVM)的直方图获得了基于本地采样的特征表示之间的最佳性能。为了提高STC识别,我们应用全局采样特征表示。通过从全局采样计算HOG描述符来生成Global Hog(GHOG)。 GHOG使得能够更好的字符结构建模并比基于本地采样的特征表示获得更好的性能。 GHOG还优于两个基准数据集中的现有方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号