Quadratic optimization fine tuning for the learning phase of SVM

1Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper presents a study of the Quadratic optimization Problem (QP) lying on the learning process of Support Vector Machines (SVM). Taking the Karush-Kuhn-Tucker (KKT) optimality conditions, we present the strategy of implementation of the SVM-QP following two classical approaches: i) active set, also divided in primal and dual spaces, methods and ii) interior point methods. We also present the general extension to treat large scale applications consisting in a general decomposition of the QP problem into smaller ones. In the same manner, we discuss some considerations to take into account to start the general learning process. We compare the performances of the optimization strategies using some well-known benchmark databases. © Springer-Verlag Berlin Heidelberg 2005.

Cite

CITATION STYLE

APA

González-Mendoza, M., Hernández-Gress, N., & Titli, A. (2005). Quadratic optimization fine tuning for the learning phase of SVM. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 3563 LNCS, pp. 347–357). Springer Verlag. https://doi.org/10.1007/11533962_31

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free