Computational properties of probabilistic neural networks

2Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We discuss the problem of overfitting of probabilistic neural networks in the framework of statistical pattern recognition. The probabilistic approach to neural networks provides a statistically justified subspace method of classification. The underlying structural mixture model includes binary structural parameters and can be optimized by EM algorithm in full generality. Formally, the structural model reduces the number of parameters included and therefore the structural mixtures become less complex and less prone to overfitting. We illustrate how recognition accuracy and the effect of overfitting is influenced by mixture complexity and by the size of training data set. © 2010 Springer-Verlag Berlin Heidelberg.

Cite

CITATION STYLE

APA

Grim, J., & Hora, J. (2010). Computational properties of probabilistic neural networks. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6354 LNCS, pp. 31–40). https://doi.org/10.1007/978-3-642-15825-4_4

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free