Approximate reduction from AUC maximization to 1-norm soft margin optimization

2Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Finding linear classifiers that maximize AUC scores is important in ranking research. This is naturally formulated as a 1-norm hard/soft margin optimization problem over pn pairs of p positive and n negative instances. However, directly solving the optimization problems is impractical since the problem size (pn) is quadratically larger than the given sample size (p+n). In this paper, we give (approximate) reductions from the problems to hard/soft margin optimization problems of linear size. First, for the hard margin case, we show that the problem is reduced to a hard margin optimization problem over p+n instances in which the bias constant term is to be optimized. Then, for the soft margin case, we show that the problem is approximately reduced to a soft margin optimization problem over p+n instances for which the resulting linear classifier is guaranteed to have a certain margin over pairs. © 2011 Springer-Verlag.

Cite

CITATION STYLE

APA

Suehiro, D., Hatano, K., & Takimoto, E. (2011). Approximate reduction from AUC maximization to 1-norm soft margin optimization. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6925 LNAI, pp. 324–337). https://doi.org/10.1007/978-3-642-24412-4_26

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free