4.5 Article

A New Attention-Based LSTM for Image Captioning

Journal

NEURAL PROCESSING LETTERS
Volume 54, Issue 4, Pages 3157-3171

Publisher

SPRINGER
DOI: 10.1007/s11063-022-10759-z

Keywords

Image caption; Attention; Long short-term memory; Deep learning

Funding

  1. National Science and Technology Major Project [2020YFA0713504]
  2. CERNET Innovation Project [NGII20180309]
  3. Scientific Research Fund of Hunan Provincial Education Department [210153]

Ask authors/readers for more resources

This paper proposes an attentional LSTM (ALSTM) for image captioning. Unlike traditional LSTM, ALSTM can refine input vector by learning from network hidden states and sequential context information. ALSTM is used as the decoder in some classical frameworks and demonstrates how to obtain effective visual/context attention. Extensive evaluations show the superiority of ALSTM in generating high-quality image descriptions.
Image captioning aims to describe the content of an image with a complete and natural sentence. Recently, the image captioning methods with encoder-decoder architecture has made great progress, in which LSTM became a dominant decoder to generate word sequence. However, in the decoder stage, the input vector keep same and there is much uncorrelated with previously visual parts or generated words. In this paper, we propose an attentional LSTM (ALSTM) and show how to integrate it within state-of-the-art automatic image captioning framework. Instead of traditional LSTM in existing models, ALSTM learns to refine input vector from network hidden states and sequential context information. Thus ALSTM can attend more relevant features such as spatial attention, visual relations and pay more attention on the most relevant context words. Moreover, ALSTM is utilized as the decoder in some classical frameworks and shows how to get effective visual/context attention to update input vector. Extensive quantitative and qualitative evaluations on the Flickr30K and MSCOCO image datasets with modified network illustrate the superiority of ALSTM. ALSTM based methods can generate high quality descriptions by combining sequence context and relations.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available