For neural networks, learning from dichotomous random samples is difficult. An example is learning of a Bayesian discriminant function. However, one-hidden-layer neural networks with fewer inner parameters can learn from such signals better than ordinary ones. We show that such neural networks can be used for approximating multi-category Bayesian discriminant functions when the state-conditional probability distributions are two dimensional normal distributions. Results of a simple simulation are shown as examples. © Springer-Verlag Berlin Heidelberg 2008.
CITATION STYLE
Ito, Y., Srinivasan, C., & Izumi, H. (2008). Multi-category bayesian decision by neural networks. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 5163 LNCS, pp. 21–30). https://doi.org/10.1007/978-3-540-87536-9_3
Mendeley helps you to discover research relevant for your work.