3.8 Proceedings Paper

Text Recognition - Real World Data and Where to Find Them

出版社

IEEE COMPUTER SOC
DOI: 10.1109/ICPR48806.2021.9412868

关键词

-

资金

  1. Czech Technical University [SGS20/171/OHK3/3T/13]
  2. MEYS VVV project [CZ.02.1.01/0.0/0.0/16 019/0000765]
  3. Spanish Research project [TIN2017-89779-P]
  4. CERCA Programme/Generalitat de Catalunya

向作者/读者索取更多资源

The method proposed leverages weakly annotated images to enhance text extraction pipelines, by combining imprecise text transcriptions with weak annotations to generate nearly error-free instances of scene text for training, resulting in consistent improvements in accuracy for state-of-the-art recognition models.
We present a method for exploiting weakly annotated images to improve text extraction pipelines. The approach uses an arbitrary end-to-end text recognition system to obtain text region proposals and their, possibly erroneous, transcriptions. The method includes matching of imprecise transcriptions to weak annotations and an edit distance guided neighbourhood search. It produces nearly error-free, localised instances of scene text, which we treat as pseudo ground truth (PGT). The method is applied to two weakly-annotated datasets. Training with the extracted PGT consistently improves the accuracy of a state of the art recognition model, by 3.7% on average, across different benchmark datasets (image domains) and 24.5% on one of the weakly annotated datasets.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据