4.5 Article

How effective is the Grey Wolf optimizer in training multi-layer perceptrons

Journal

APPLIED INTELLIGENCE
Volume 43, Issue 1, Pages 150-161

Publisher

SPRINGER
DOI: 10.1007/s10489-014-0645-7

Keywords

Grey Wolf optimizer; MLP; Learning neural network; Evolutionary algorithm; Multi-layer perceptron

Ask authors/readers for more resources

This paper employs the recently proposed Grey Wolf Optimizer (GWO) for training Multi-Layer Perceptron (MLP) for the first time. Eight standard datasets including five classification and three function-approximation datasets are utilized to benchmark the performance of the proposed method. For verification, the results are compared with some of the most well-known evolutionary trainers: Particle Swarm Optimization (PSO), Genetic Algorithm (GA), Ant Colony Optimization (ACO), Evolution Strategy (ES), and Population-based Incremental Learning (PBIL). The statistical results prove the GWO algorithm is able to provide very competitive results in terms of improved local optima avoidance. The results also demonstrate a high level of accuracy in classification and approximation of the proposed trainer.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available