4.7 Article

Neural architecture search with interpretable meta-features and fast predictors

Journal

INFORMATION SCIENCES
Volume 649, Issue -, Pages -

Publisher

ELSEVIER SCIENCE INC
DOI: 10.1016/j.ins.2023.119642

Keywords

Neural architecture search; Meta-learning; Prediction-based NAS; Interpretability; Image classification

Ask authors/readers for more resources

This paper proposes a Prediction-based and interpretable Meta-Learning method called MbML-NAS, which can generalize to different search spaces and datasets using less data. The method uses interpretable meta-features extracted from neural architectures and regression models as meta-predictors to infer Convolutional Networks performances.
Neural Architecture Search (NAS) is well-known for automatizing neural architecture design and finding better architectures. Although NAS methods have shown substantial progress over the years, most still suffer from data inefficiency, high model complexity, and lack of interpretability. This paper proposes a solution for these problems by introducing a Prediction-based and interpretable Meta-Learning method called MbML-NAS, capable of generalizing to different search spaces and datasets using less data than several state-of-the-art NAS methods. The proposal uses interpretable meta-features extracted from neural architectures and regression models as meta-predictors to infer Convolutional Networks performances. Experiments compared MbMLNAS with a graph-based Neural Predictor, state-of-the-art NAS methods, a lower bound, and an upper bound Oracle baseline. Furthermore, an interpretability analysis of meta-features and meta-predictors is presented. As a result, using at least 172 examples representing 0.04% and 1.1% of the popular NAS-Bench-101 and NAS-Bench-201 search spaces, MbML-NAS find architectures with better or comparable performances than most baselines, including the Oracle. Moreover, the results showed the potential for using simple meta-features to generalize across NAS search spaces and datasets, encoding neural architectures so that even linear models can accurately predict their performances. Additionally, novel meta-datasets suitable for NAS are proposed to facilitate research on NAS.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available