Measuring Instance Hardness Using Data Complexity Measures

7Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Assessing the hardness of each instance in a problem is an important meta-knowledge which may leverage advances in Machine Learning. In classification problems, an instance can be regarded as difficult if it gets systematically misclassified by a diverse set of classification techniques with different biases. The instance hardness measures were proposed with the aim of relating data characteristics to this notion of intrinsic difficulty of the instances. There are also in the literature a large set of measures which are dedicated at describing the difficulty of a classification problem from a dataset-level perspective. In this paper these measures are decomposed at the instance-level, giving a perspective of how each individual example in a dataset contributes to its overall complexity. Experiments on synthetic and benchmark datasets demonstrate the proposed measures can provide a complementary instance hardness perspective when compared to those from related literature.

Cite

CITATION STYLE

APA

Arruda, J. L. M., Prudêncio, R. B. C., & Lorena, A. C. (2020). Measuring Instance Hardness Using Data Complexity Measures. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12320 LNAI, pp. 483–497). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-61380-8_33

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free