Journal
APPLIED INTELLIGENCE
Volume 43, Issue 1, Pages 150-161Publisher
SPRINGER
DOI: 10.1007/s10489-014-0645-7
Keywords
Grey Wolf optimizer; MLP; Learning neural network; Evolutionary algorithm; Multi-layer perceptron
Categories
Ask authors/readers for more resources
This paper employs the recently proposed Grey Wolf Optimizer (GWO) for training Multi-Layer Perceptron (MLP) for the first time. Eight standard datasets including five classification and three function-approximation datasets are utilized to benchmark the performance of the proposed method. For verification, the results are compared with some of the most well-known evolutionary trainers: Particle Swarm Optimization (PSO), Genetic Algorithm (GA), Ant Colony Optimization (ACO), Evolution Strategy (ES), and Population-based Incremental Learning (PBIL). The statistical results prove the GWO algorithm is able to provide very competitive results in terms of improved local optima avoidance. The results also demonstrate a high level of accuracy in classification and approximation of the proposed trainer.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available