An evolutionary multi-objective neural network optimizer with bias-based pruning heuristic

4Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Neural network design aims for high classification accuracy and low network architecture complexity. It is also known that simultaneous optimization of both model accuracy and complexity improves generalization while avoiding overfitting on data. We describe a neural network training procedure that uses multi-objective optimization to evolve networks which are optimal both with respect to classification accuracy and architectural complexity. The NSGA-II algorithm is employed to evolve a population of neural networks that are minimal in both training error and a Minimum Description Length-based network complexity measure. We further propose a pruning rule based on the following heuristic: connections to or from a node may be severed if their weight values are smaller than the network's smallest bias. Experiments on benchmark datasets show that the proposed evolutionary multi-objective approach to neural network design employing the bias-based pruning heuristic yields networks that have far fewer connections without seriously compromising generalization performance when compared to other existing evolutionary optimization algorithms. © Springer-Verlag Berlin Heidelberg 2007.

Cite

CITATION STYLE

APA

Naval, P. C., & Yusiong, J. P. T. (2007). An evolutionary multi-objective neural network optimizer with bias-based pruning heuristic. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 4493 LNCS, pp. 174–183). Springer Verlag. https://doi.org/10.1007/978-3-540-72395-0_23

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free