Evaluating reliability of single classifications of neural networks

2Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Current machine learning algorithms perform well on many problem domains, but in risk-sensitive decision making, for example in medicine and finance, common evaluation methods that give overall assessments of models fail to gain trust among experts, as they do not provide any information about single predictions. We continue the previous work on approaches for evaluating the reliability of single classifications where we focus on methods that are model independent. These methods have been shown to be successful in their narrow fields of application, so we constructed a testing methodology to evaluate these methods in straightforward, general-use test cases. For the evaluation, we had to derive a statistical reference function, which enables comparison between the reliability estimators and the model's own predictions. We compare five different approaches and evaluate them on a simple neural network with several artificial and real-world domains. The results indicate that reliability estimators CNK and LCV can be used to improve the model's predictions. © 2011 Springer-Verlag.

Cite

CITATION STYLE

APA

Pevec, D., Štrumbelj, E., & Kononenko, I. (2011). Evaluating reliability of single classifications of neural networks. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6593 LNCS, pp. 22–30). https://doi.org/10.1007/978-3-642-20282-7_3

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free