3.8 Proceedings Paper

Adapted GooLeNet for Visual Question Answering

Publisher

IEEE
DOI: 10.1109/ICMCCE.2018.00132

Keywords

visual question answering; Adapted GooLeNet; MUTAN

Funding

  1. National Natural Science Foundation of China [61473300]

Ask authors/readers for more resources

Visual Question Answering (VQA) aims at answering a question about an image. In this work, we introduce an effective architecture --Adapted GooLeNet (AG)-- into a typical VQA method MUTAN instead of LSTM for question features capturing. This improvement can capture more levels of language granularities in parallel, because of the various sizes of filters in AG. The empirical study on the benchmark dataset of VQA demonstrates that capturing sentence features on different levels of granularities benefit sentence modelling by utilizing AG.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available