Synthetic Data for Text Localisation in Natural Images

2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)(2016)

引用 1662|浏览188
暂无评分
摘要
In this paper we introduce a new method for text detection in natural images. The method comprises two contributions: First, a fast and scalable engine to generate synthetic images of text in clutter. This engine overlays synthetic text to existing background images in a natural way, accounting for the local 3D scene geometry. Second, we use the synthetic images to train a Fully-Convolutional Regression Network (FCRN) which efficiently performs text detection and bounding-box regression at all locations and multiple scales in an image. We discuss the relation of FCRN to the recently-introduced YOLO detector, as well as other end-to-end object detection systems based on deep learning. The resulting detection network significantly out performs current methods for text detection in natural images, achieving an F-measure of 84.2% on the standard ICDAR 2013 benchmark. Furthermore, it can process 15 images per second on a GPU.
更多
查看译文
关键词
text localisation,synthetic image generation,synthetic text,background images,local 3D scene geometry,fully-convolutional regression network training,FCRN training,text detection,bounding-box regression,FCRN,YOLO detector,end-to-end object detection systems,deep learning,natural images,F-measure,standard ICDAR 2013 benchmark
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要