期刊
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021
卷 -, 期 -, 页码 12021-12030出版社
IEEE COMPUTER SOC
DOI: 10.1109/CVPR46437.2021.01185
关键词
-
资金
- STCSM Projects [20511100400, 20511102702]
- Shanghai Municipal Science and Technology Major Projects [2017SHZDZX01, 2018SHZDZX01]
- Shanghai Research and Innovation Functional Program [17DZ2260900]
- Program for Professor of Special Appointment (Eastern Scholar) at Shanghai Institutions of Higher Learning
- ZJLab
The study introduces a text-focused super-resolution framework, which utilizes Transformer and self-attention module to extract sequential information, and position-aware and content-aware modules to emphasize character position and content. Weighted cross-entropy loss is used to address the issue of indistinguishable characters in low-resolution conditions.
Image super-resolution, which is often regarded as a preprocessing procedure of scene text recognition, aims to recover the realistic features from a low-resolution text image. It has always been challenging due to large variations in text shapes, fonts, backgrounds, etc. However, most existing methods employ generic super-resolution frameworks to handle scene text images while ignoring text-specific properties such as text-level layouts and character-level details. In this paper, we establish a text-focused super-resolution framework, called Scene Text Telescope (STT). In terms of text-level layouts, we propose a Transformer-Based Super-Resolution Network (TBSRN) containing a Self-Attention Module to extract sequential information, which is robust to tackle the texts in arbitrary orientations. In terms of character-level details, we propose a Position-Aware Module and a Content-Aware Module to highlight the position and the content of each character. By observing that some characters look indistinguishable in low-resolution conditions, we use a weighted cross-entropy loss to tackle this problem. We conduct extensive experiments, including text recognition with pre-trained recognizers and image quality evaluation, on TextZoom and several scene text recognition benchmarks to assess the super-resolution images. The experimental results show that our STT can indeed generate text-focused super-resolution images and outperform the existing methods in terms of recognition accuracy.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据