Nonlinear Dimension Reduction:

Post on 31-Jan-2016

36 views 0 download

Tags:

description

Nonlinear Dimension Reduction:. Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao. Outline. Nonlinear Dimension Reduction Semi-Definite Embedding Local Linear Embedding Experiments. Dimension Reduction. - PowerPoint PPT Presentation

transcript

Nonlinear Dimension Reduction:

Semi-Definite Embedding vs. Local Linear Embedding

Li Zhang and Lin Liao

Outline

Nonlinear Dimension Reduction Semi-Definite Embedding Local Linear Embedding Experiments

Dimension Reduction To understand images in terms of their

basic modes of variability. Unsupervised learning problem: Given

N high dimensional input Xi RD, find a faithful one-to-one mapping to N low dimensional output Yi Rd and d<D.

Methods: Linear methods (PCA, MDS): subspace Nonlinear methods (SDE, LLE): manifold

Semi-Definite Embedding

Given input X=(X1,...,XN) and k Find the k nearest neighbors for each

input Xi

Formulate and solve a corresponding semi-definite programming problem; find optimal Gram matrix of output K=YTY

Extract approximately a low dimensional embedding Y from the eigenvectors and eigenvalues of Gram matrix K

Semi-Definite Programming

Maximize C·XSubject to

AX=bmatrix(X) is positive semi-

definitewhere X is a vector with size n2, and

matrix(X) is a n by n matrix reshaped from X

Semi-Definite Programming

Constraints: Maintain the distance between neighbors

|Yi-Yj|2=|Xi-Xj|2for each pair of neighbor (i,j)

Kii+Kjj-Kij-Kji= Gii+Gjj-Gij-Gji where K=YTY,G=XTX

Constrain the output centered on the originΣYi=0 ΣKij=0

K is positive semidefinite

Semi-Definite Programming

Objective function Maximize the sum of pairwise

squared distance between outputsΣij|Yi-Yj|2 Tr(K)

Semi-Definite Programming

Solve the best K using any SDP solver CSDP (fast, stable) SeDuMi (stable, slow) SDPT3 (new, fastest, not well tested)

Locally Linear Embedding

Swiss Roll

N=800

SDE, k=4

LLE, k=18

LLE on Swiss Roll, varying K

K=5 K=6

K=8 K=10

LLE on Swiss Roll, varying K

K=12 K=14

K=16 K=18

LLE on Swiss Roll, varying K

K=20 K=30

K=40 K=60

Twos

N=638

SDE, k=4

LLE, k=18

Teapots

N=400

SDE, k=4

LLE, k=12

LLE on Teapot, varying N

N=400 N=200

N=100 N=50

Faces

N=1900

SDE, failed

LLE, k=12

SDE versus LLE

Similar idea First, compute neighborhoods in the

input space Second, construct a square matrix to

characterize local relationship between input data.

Finally, compute low-dimension embedding using the eigenvectors of the matrix

SDE versus LLE

Different performance SDE: good quality, more robust to

sparse samples, but optimization is slow and hard to scale to large data set

LLE: fast, scalable to large data set, but low quality when samples are sparse, due to locally linear assumption