4.7 Article

Fast and Unsupervised Neural Architecture Evolution for Visual Representation Learning

期刊

IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE
卷 16, 期 3, 页码 22-32

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/MCI.2021.3084394

关键词

-

资金

  1. Key Research and Development Program of Shandong Province [2019JZZY011101]
  2. [62076016]
  3. [61876015]

向作者/读者索取更多资源

This paper proposes a fast and unsupervised neural architecture evolution method that accelerates the evolutionary process using self-supervision, achieving state-of-the-art performance in downstream applications.
Unsupervised visual representation learning is one of the hottest topics in computer vision, yet performance still lags behind compared with the best supervised learning methods. At the same time, neural architecture search (NAS) has produced state-of-the-art results on various visual tasks. It is a natural idea to explore NAS as a way to improve unsupervised representation learning, yet it remains largely unexplored. In this paper, we propose a Fast and Unsupervised Neural Architecture Evolution (FaUNAE) method to evolve an existing architecture, manually constructed or the result of NAS on a small dataset, to a new architecture that can operate on a larger dataset.This partial optimization can utilize prior knowledge to reduce search cost and improve search efficiency. The evolution is self-supervised where the contrast loss is used as the evaluation metric in a student-teacher framework. By eliminating the inferior or least promising operations, the evolutionary process is greatly accelerated. Experimental results show that we achieve state-of-the-art performance for downstream applications, such as object recognition, object detection, and instance segmentation.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据