Local Fisher Discriminant Analysis for Supervised
Dimensionality Reduction
Presented by Xianwang WangPresented by Xianwang Wang
Masashi Sugiyama
Dimensionality Reduction Goal
Embed high-dimensional data to low-dimensional space Preserve intrinsic information
Example
High dimension
3-dimension
Categories Nonlinear
ISOMAP Locally Linear Embedding (LLE) Laplacian Eigenmap (LE)
Linear Principal Components Analysis (PCA) Locality-Preserving Projection (LPP) Fisher Discriminant Analysis (FDA)
Unsupervised S-ISOMAP, S-LLE, PCA
Supervised LPP, FDA
Formulation Number of samples:
d-dimensional samples:
Class labels :
Number of samples in the class :
Data matrix :
Embedded samples:
Goal for linear dimensionality Reduction Find a transformation matrix
Use Iris data for demos (http://archive.ics.uci.edu/ml/machine-learning-databases/iris/iris.data) Attribute Information:
sepal length in cm sepal width in cm petal length in cm petal width in cm
class: Iris Setosa; Iris Versicolour; Iris Virginica
FDA(1) Mean of samples in the class
Mean of all samples
Within-class scatter matrix
Between-class scatter matrix
FDA(2) Maximize the following objective
Maximize the following constrained optimization problem equivalently
Use the lagrangian,
Apply KKT conditions
Demo
LPP Minimize
Equivalently
We can get
Demo
Local Fisher Discriminant Analysis(LFDA) FDA can perform poorly if samples in some class form
several separate clusters
LPP can make samples of different classes overlapped if they are close in the original high dimensional space
LFDA combines the idea of FDA and LPP
LFDA(1) Reformulating FDA
LFDA(2) Definition of LFDA
LFDA(3) Maximize the following objective
Equivalently,
Similarly, we can get
Demo
Conclusion
LFDA provided more separate embedding than FDA and LPP
FDA (globally), while LFDA(locally)
More discussion about efficiently computing of LFDA transformation matrix and Kernel LFDA in the paper
Questions?