Skip to main content

On using nearly-independent feature families for high precision and confidence

19Citations
Citations of this article
54Readers
Mendeley users who have this article in their library.

This artice is free to access.

Abstract

Consider learning tasks where the precision requirement is very high, for example a 99 % precision requirement for a video classification application. We report that when very different sources of evidence such as text, audio, and video features are available, combining the outputs of base classifiers trained on each feature type separately, aka late fusion, can substantially increase the recall of the combination at high precisions, compared to the performance of a single classifier trained on all the feature types, i.e.; early fusion, or compared to the individual base classifiers. We show how the probability of a joint false-positive mistake can be less - in some cases significantly less - than the product of individual probabilities of conditional false-positive mistakes (a NoisyOR combination). Our analysis highlights a simple key criterion for this boosted precision phenomenon and justifies referring to such feature families as (nearly) independent. We assess the relevant factors for achieving high precision empirically, and explore combination techniques informed by the analysis. © 2013 The Author(s).

Cite

CITATION STYLE

APA

Madani, O., Georg, M., & Ross, D. (2013). On using nearly-independent feature families for high precision and confidence. In Machine Learning (Vol. 92, pp. 457–477). https://doi.org/10.1007/s10994-013-5377-0

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free