Neural minimax classifiers

0Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Many supervised learning algorithms are based on the assumption that the training data set reflects the underlying statistical model of the real data. However, this stationarity assumption may be partially violated in practice: for instance, if the cost of collecting data is class dependent, the class priors of the training data set may be different from that of the test set. A robust solution to this problem is selecting the classifier that minimize the error probability under the worst case conditions. This is known as the minimax strategy. In this paper we propose a mechanism to train a neural network in order to estimate the minimax classifier that is robust to changes in the class priors. This procedure is illustrated on a softmax-based neural network, although it can be applied to other structures. Several experimental results show the advantages of the proposed methods with respect to other approaches. © Springer-Verlag Berlin Heidelberg 2002.

Cite

CITATION STYLE

APA

Alaiz-Rodríguez, R., & Cid-Sueiro, J. (2002). Neural minimax classifiers. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 2415 LNCS, pp. 408–413). Springer Verlag. https://doi.org/10.1007/3-540-46084-5_66

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free