Constant rate approximate maximum margin algorithms

3Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

We present a new class of Perceptron-like algorithms with margin in which the "effective" learning rate ηeff, denned as the ratio of the learning rate to the length of the weight vector, remains constant. We prove that for ηeff sufficiently small the new algorithms converge in a finite number of steps and show that there exists a limit of the parameters involved in which convergence leads to classification with maximum margin. A soft margin extension for Perceptron-like large margin classifiers is also discussed. © Springer-Verlag Berlin Heidelberg 2006.

Cite

CITATION STYLE

APA

Tsampouka, P., & Shawe-Taylor, J. (2006). Constant rate approximate maximum margin algorithms. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 4212 LNAI, pp. 437–448). Springer Verlag. https://doi.org/10.1007/11871842_42

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free