Arenas-Garcia, J.7; Petersen, K.4; Camps-Valls, G.5; Hansen, Lars Kai6
1 Department of Applied Mathematics and Computer Science, Technical University of Denmark2 Cognitive Systems, Department of Applied Mathematics and Computer Science, Technical University of Denmark3 Universidad Carlos III de Madrid4 SAS Institute Inc.5 Universitat de Valencia6 Copenhagen Center for Health Technology, Center, Technical University of Denmark7 Universidad Carlos III de Madrid
Feature extraction and dimensionality reduction are important tasks in many fields of science dealing with signal processing and analysis. The relevance of these techniques is increasing as current sensory devices are developed with ever higher resolution, and problems involving multimodal data sources become more common. A plethora of feature extraction methods are available in the literature collectively grouped under the field of multivariate analysis (MVA). This article provides a uniform treatment of several methods: principal component analysis (PCA), partial least squares (PLS), canonical correlation analysis (CCA), and orthonormalized PLS (OPLS), as well as their nonlinear extensions derived by means of the theory of reproducing kernel Hilbert spaces (RKHSs). We also review their connections to other methods for classification and statistical dependence estimation and introduce some recent developments to deal with the extreme cases of large-scale and low-sized problems. To illustrate the wide applicability of these methods in both classification and regression problems, we analyze their performance in a benchmark of publicly available data sets and pay special attention to specific real applications involving audio processing for music genre prediction and hyperspectral satellite image processing for Earth and climate monitoring.
I E E E - Signal Processing Magazine, 2013, Vol 30, Issue 4, p. 16-29