4.6 Article

Maximum Response Deep Learning Using Markov, Retinal & Primitive Patch Binding With GoogLeNet & VGG-19 for Large Image Retrieval

期刊

IEEE ACCESS
卷 9, 期 -, 页码 41934-41957

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2021.3063545

关键词

Feature extraction; Image color analysis; Image retrieval; Shape; Visualization; Convolutional neural networks; Histograms; Bag of words; cascade sampling; content based image retrieval; color components; maximum response for texture pattern; combination of features

资金

  1. Basic Science Research Program through the National Research Foundation of Korea (NRF) - Ministry of Education [NRF-2019R1A2C1006159]
  2. MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program [IITP-2020-2016-0-00313]

向作者/读者索取更多资源

This article presents a method of retrieving visual features from flexible image datasets and obtaining accurate results through a combination of deep convolutional neural networks and content-based image retrieval schemes.
Smart and productive image retrieval from flexible image datasets is an unavoidable necessity of the current period. Crude picture marks are imperative to mirror the visual ascribes for content-based image retrieval (CBIR). Algorithmically enlightened and recognized visual substance structure image marks to accurately file and recover comparative outcomes. Consequently, highlighted vectors ought to contain adequate image data with color, shape, objects, spatial data viewpoints to recognize image class as a qualifying applicant. This article presents the maximum response of visual features of an image over profound convolutional neural networks in blend with an innovative content-based image retrieval plan to recover phenomenally precise outcomes. For this determination, a serial fusion of GoogLeNet and VGG-19 based generated signatures are formulated with visual features including texture, color and shape. Initially, the maximum response is calculated for texture pattern by using Markov Random Field (MRF) classifier. Thereafter, cascaded samples are passed through a human retinal system like descriptor named Fast Retina Keypoint (FREAK) for corresponding fundamental points through the image. GoogLeNet and VGG-19 are applied to extract deep features of an image; hence color components are obtained using a correlogram. Finally, all the image signatures are combined and passed through the BoW scheme. The proposed method is applied experimentally to challenging datasets, including Caltech-256, ALOT (250), Corel 1000, Cifar-100, and Cifar 10. Remarkable precision,Recall and F-score results obtained.The texture dataset ALOT (250) with the uppermost precision rate 0.99 for a maximum of its categories, whereas Caltech-256 gives 0.66 precision, and Corel 1000 0.99 for VGG-19 and 0.95 for GoogLeNet. Recall, F-score, ARR and ARP rates shows the significant rates in most of the image categories.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据