Correcting binary imprecise classifiers: Local vs global approach

3Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper proposes a simple strategy for combining binary classifiers with imprecise probabilities as outputs. Our combination strategy consists in computing a set of probability distributions by solving an optimization problem whose constraints depend on the classifiers outputs. However, the classifiers may provide assessments that are jointly incoherent, in which case the set of probability distributions satisfying all the constraints is empty. We study different correction strategies for restoring this consistency, by relaxing the constraints of the optimization problem so that it becomes feasible. In particular, we propose and compare a global strategy, where all constraints are relaxed to the same level, to a local strategy, where some constraints may be relaxed more than others. The local discounting strategy proves to give very good results compared both to single classifier approaches and to classifier combination schemes using a global correction scheme. © 2012 Springer-Verlag.

Cite

CITATION STYLE

APA

Destercke, S., & Quost, B. (2012). Correcting binary imprecise classifiers: Local vs global approach. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7520 LNAI, pp. 299–310). https://doi.org/10.1007/978-3-642-33362-0_23

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free