An Improved Training Algorithm for Nonlinear Kernel Discriminants

19Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

A simple method to derive nonlinear discriminants is to map the samples into a high-dimensional feature space F using a nonlinear function and then to perform a linear discriminant analysis in F. Clearly, if F is a very high, or even infinitely, dimensional space, designing such a receiver may be a computationally intractable problem. However, using Mercer kernels, this problem can be solved without explicitly mapping the data to F. Recently, a powerful method of obtaining nonlinear kernel Fisher discriminants (KFDs) has been proposed, and very promising results were reported when compared with the other state-of-the-art classification techniques. In this paper, we present an extension of the KFD method that is also based on Mercer kernels. Our approach, which is called the nonlinear kernel second-order discriminant (KSOD), consists of determining a nonlinear receiver via optimization of a general form of second-order measures of performance. We also propose a complexity control procedure in order to improve the performance of these classifiers when few training data are available. Finally, simulations compare our approach with the KFD method. © 2004, IEEE. All rights reserved.

Cite

CITATION STYLE

APA

Abdallah, F., Richard, C., & Lengellé, R. (2004). An Improved Training Algorithm for Nonlinear Kernel Discriminants. IEEE Transactions on Signal Processing, 52(10), 2798–2806. https://doi.org/10.1109/TSP.2004.834346

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free