Deriving consensus rankings from benchmarking experiments

18Citations
Citations of this article
27Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Whereas benchmarking experiments are very frequently used to investigate the performance of statistical or machine learning algorithms for supervised and unsupervised learning tasks, overall analyses of such experiments are typically only carried out on a heuristic basis, if at all. We suggest to determine winners, and more generally, to derive a consensus ranking of the algorithms, as the linear order on the algorithms which minimizes average symmetric distance (Kemeny-Snell distance) to the performance relations on the individual benchmark data sets. This leads to binary programming problems which can typically be solved reasonably efficiently. We apply the approach to a medium-scale benchmarking experiment to assess the performance of Support Vector Machines in regression and classification problems, and compare the obtained consensus ranking with rankings obtained by simple scoring and Bradley-Terry modeling.

Cite

CITATION STYLE

APA

Hornik, K., & Meyer, D. (2007). Deriving consensus rankings from benchmarking experiments. In Studies in Classification, Data Analysis, and Knowledge Organization (pp. 163–170). Kluwer Academic Publishers. https://doi.org/10.1007/978-3-540-70981-7_19

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free