3.8 Proceedings Paper

The Image Annotation Method by Convolutional Features from Intermediate Layer of Deep Learning Based on Internet of Things

Publisher

IEEE
DOI: 10.1109/MSN48538.2019.00066

Keywords

deep learning; image annotation; convolutional result; positive mean vector; eigenvector; internet of things

Funding

  1. National Natural Science Foundation of China [61972056, 61981340416]
  2. Open Research Fund of Hunan Provincial Key Laboratory of Intelligent Processing of Big Data on Transportation [2015TP1005]
  3. Changsha Science and Technology Planning [KQ1703018, KQ1706064, KQ1703018-01, KQ1703018-04]
  4. Research Foundation of Education Bureau of Hunan Province [17A007]
  5. Changsha Industrial Science and Technology Commissioner [2017-7]
  6. Junior Faculty Development Program Project of Changsha University of Science and Technology [2019QJCZ011]

Ask authors/readers for more resources

Existing image annotation methods that employ convolutional features of deep learning methods from the Internet of Things (IoT) have a number of limitations, including complex training and high space/time expenses associated with the image annotation procedure. Accordingly, this paper proposes an innovative method in which the visual features of the image are presented by the intermediate layer features of deep learning, while semantic concepts are represented by mean vectors of positive samples. Firstly, the convolutional result is directly output in the form of low-level visual features through the mid-level of the pre-trained deep learning model, with the image being represented by sparse coding in the IoT. Secondly, the positive mean vector method is used to construct visual feature vectors for each text vocabulary item, so that a visual feature vector database is created. Finally, the visual feature vector similarity between the testing image and all text vocabulary is calculated, and the vocabulary with the largest similarity is taken from the IoT as the words used for annotation. Experiments on multiple datasets demonstrate the effectiveness of the proposed method; in terms of F1 score, the proposed method's performance on the Corel5k and IAPR TC-12 datasets is superior to that of MBRM, JEC-AF, JEC-DF, and 2PKNN with end-to-end deep features.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available