PLS and dimension reduction for classification

49Citations
Citations of this article
68Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Barker and Rayens (J Chemometrics 17:166-173, 2003) offered convincing arguments that partial least squares (PLS) is to be preferred over principal components analysis (PCA) when discrimination is the goal and dimension reduction is required, since at least with PLS as the dimension reduction tool, information involving group separation is directly involved in the structure extraction. In this paper the basic results in Barker and Rayens (J Chemometrics 17:166-173, 2003) are reviewed and some of their ideas and comparisons are illustrated on a real data set, something which Barker and Rayens did not do. More importantly, new results are introduced, including a formal proof for the superiority of PLS over PCA in the two-group case, as well as new connections between PLS for discrimination and an extended class of PLS-like techniques known as "oriented PLS" (OrPLS). In the latter case, a particularly simple subclass of OrPLS procedures, when used to achieve the dimension reduction, is shown to always produce a lower misclassification rate than when "ordinary" PLS is used for the same purpose. © Springer-Verlag 2007.

Cite

CITATION STYLE

APA

Liu, Y., & Rayens, W. (2007). PLS and dimension reduction for classification. Computational Statistics, 22(2), 189–208. https://doi.org/10.1007/s00180-007-0039-y

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free