4.6 Article

Enhanced Feature Model Based Hybrid Neural Network for Text Detection on Signboard, Billboard and News Tickers

Journal

IEEE ACCESS
Volume 11, Issue -, Pages 41524-41534

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2023.3264569

Keywords

Feature extraction; Text recognition; Character recognition; Videos; Convolutional neural networks; Recurrent neural networks; Optical character recognition; Image analysis; Deep learning; Natural scene recognition; machine learning; deep learning; image processing; natural scene Tamil character recognition

Ask authors/readers for more resources

Recognizing text from nature scene images and videos is a challenging task due to their complex features and variations. However, text recognition is highly useful in various applications. This research paper proposes a model that combines Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN) to successfully detect and recognize text characters in natural images. The experimental results demonstrate that the proposed model outperforms other methods on multiple datasets.
Recognizing text from the nature scene images and videos has been the challenging task of computer vision and machine learning research community in recent years. These texts are difficult to recognize because of their shapes, complex backgrounds, color, shape and size variations. However, text recognition is very much useful in indexing, keyword-based image and video search, and information retrieval. In this research paper, a model is proposed to detect the isolated text characters in the photographic images of natural scenes. The proposed model uses the combination of Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN) for recognizing the text in natural images. The model uses two networks, where the first network combines the low-level and middle-level features to increase the feature size and passes the enriched information to the second network. Here, features are again widened by combining with high-level features, resulting in powerful and robust features. To evaluate the proposed model, ICDAR2003 (IC03), ICDAR2013 (IC13), SVT (Street View Text) datasets have been used. And an extensive Tamil news tickers image dataset has been developed to evaluate the model. The experimental results show that the combined feature fusion technique outperforms the other methods on the ICDAR2003, ICDAR2013, SVT and Tamil news tickers datasets.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available