3.8 Proceedings Paper

HCADecoder: A Hybrid CTC-Attention Decoder for Chinese Text Recognition

Journal

DOCUMENT ANALYSIS AND RECOGNITION, ICDAR 2021, PT III
Volume 12823, Issue -, Pages 172-187

Publisher

SPRINGER INTERNATIONAL PUBLISHING AG
DOI: 10.1007/978-3-030-86334-0_12

Keywords

Chinese text recognition; CTC-Attention; Subword

Funding

  1. National Natural Science Foundation of China [U2034211, 62006017]
  2. Fundamental Research Funds for the Central Universities [2020JBZD010]
  3. Beijing Natural Science Foundation [L191016]

Ask authors/readers for more resources

This study proposes a hybrid CTC-Attention decoder for Chinese text recognition based on the characteristic of Chinese word frequency distribution. Experimental results demonstrate the effectiveness of the proposed method, especially for long texts. The code will be publicly available on GitHub.
Text recognition has attracted much attention and achieved exciting results on several commonly used public English datasets in recent years. However, most of these well-established methods, such as connectionist temporal classification (CTC)-based methods and attention-based methods, pay less attention to challenges on the Chinese scene, especially for long text sequences. In this paper, we exploit the characteristic of Chinese word frequency distribution and propose a hybrid CTC-Attention decoder (HCADecoder) supervised with bigram mixture labels for Chinese text recognition. Specifically, we first add high-frequency bigram subwords into the original unigram labels to construct the mixture bigram label, which can shorten the decoding length. Then, in the decoding stage, the CTC module outputs a preliminary result, in which confused predictions are replaced with bigram subwords. The attention module utilizes the preliminary result and outputs the final result. Experimental results on four Chinese datasets demonstrate the effectiveness of the proposed method for Chinese text recognition, especially for long texts. Code will be made publicly available(https://github.com/lukecsq/hybrid-CTC-Attention).

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available