4.7 Article

LoNAS: Low-Cost Neural Architecture Search Using a Three-Stage Evolutionary Algorithm

期刊

IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE
卷 18, 期 2, 页码 78-93

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/MCI.2023.3245799

关键词

Costs; Graphics processing units; Computer architecture; Evolutionary computation; Manuals; Network architecture; Search problems

向作者/读者索取更多资源

This paper proposes a low-cost neural architecture search method (LoNAS) that addresses the problems of existing approaches by designing a variable-architecture encoding strategy, a training-free proxy, and a three-stage evolutionary algorithm. The experimental results show that LoNAS finds network architectures with competitive performance in test accuracy and the number of parameters, using less search time and fewer computational resources.
Neural architecture search (NAS) has been widely studied to design high-performance network architectures automatically. However, existing approaches require more search time and substantial resource consumption due to their intensive architecture evaluations. Moreover, recently developed NAS algorithms are noncompetitive when combining multiple competing and conflicting objectives, e.g., the test accuracy and the number of parameters. In this paper, a low-cost NAS (LoNAS) method is proposed to address these problems. First, a variable-architecture encoding strategy based on a novel Reg Block is designed to construct high accuracy network architectures with few parameters. Second, a training-free proxy based on the neural tangent kernel (NTK) is proposed to accelerate the search process efficiently. Finally, a three-stage evolutionary algorithm (EA) based on multiple-criteria environmental selection and a set of block-based mutation operators are designed to balance exploration and exploitation better. The experimental results show that LoNAS finds network architectures with competitive performance compared to the state-of-the-art architectures in test accuracy and the number of parameters. Moreover, LoNAS uses less search time and fewer computational resources, consuming only 0.02 GPU Days with one GPU on CIFAR-10 and CIFAR-100. Furthermore, the architectures found by LoNAS on CIFAR-10 and CIFAR-100 exhibit good transferability to ImageNet-16-120, with the test accuracy surpassing that of the state-of-the-art network architectures.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据