首页> 中文期刊> 《中国科学》 >SynthText3D:synthesizing scene text images from 3D virtual worlds

SynthText3D:synthesizing scene text images from 3D virtual worlds

         

摘要

With the development of deep neural networks, the demand for a significant amount of annotated training data becomes the performance bottlenecks in many fields of research and applications. Image synthesis can generate annotated images automatically and freely, which gains increasing attention recently.In this paper, we propose to synthesize scene text images from the 3D virtual worlds, where the precise descriptions of scenes, editable illumination/visibility, and realistic physics are provided. Different from the previous methods which paste the rendered text on static 2D images, our method can render the 3D virtual scene and text instances as an entirety. In this way, real-world variations, including complex perspective transformations, various illuminations, and occlusions, can be realized in our synthesized scene text images.Moreover, the same text instances with various viewpoints can be produced by randomly moving and rotating the virtual camera, which acts as human eyes. The experiments on the standard scene text detection benchmarks using the generated synthetic data demonstrate the effectiveness and superiority of the proposed method.

著录项

  • 来源
    《中国科学》 |2020年第2期|65-78|共14页
  • 作者单位

    1. School of Electronic Information and Communications;

    Huazhong University of Science and Technology 2. School of Electronics Engineering and Computer Science;

    Peking University 3. School of Economics;

    Peking University 4. MEGVII;

  • 原文格式 PDF
  • 正文语种 chi
  • 中图分类 TP391.41;
  • 关键词

相似文献

  • 中文文献
  • 外文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号