19
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao

Nonlinear Dimension Reduction:

  • Upload
    homer

  • View
    31

  • Download
    0

Embed Size (px)

DESCRIPTION

Nonlinear Dimension Reduction:. Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao. Outline. Nonlinear Dimension Reduction Semi-Definite Embedding Local Linear Embedding Experiments. Dimension Reduction. - PowerPoint PPT Presentation

Citation preview

Page 1: Nonlinear Dimension Reduction:

Nonlinear Dimension Reduction:

Semi-Definite Embedding vs. Local Linear Embedding

Li Zhang and Lin Liao

Page 2: Nonlinear Dimension Reduction:

Outline

Nonlinear Dimension Reduction Semi-Definite Embedding Local Linear Embedding Experiments

Page 3: Nonlinear Dimension Reduction:

Dimension Reduction To understand images in terms of their

basic modes of variability. Unsupervised learning problem: Given

N high dimensional input Xi RD, find a faithful one-to-one mapping to N low dimensional output Yi Rd and d<D.

Methods: Linear methods (PCA, MDS): subspace Nonlinear methods (SDE, LLE): manifold

Page 4: Nonlinear Dimension Reduction:

Semi-Definite Embedding

Given input X=(X1,...,XN) and k Finding the k nearest neighbors for each

input Xi

Formulate and solve a corresponding semi-definite programming problem; find optimal Gram matrix of output K=YTY

Extract approximately a low dimensional embedding Y from the eigenvectors and eigenvalues of Gram matrix K

Page 5: Nonlinear Dimension Reduction:

Semi-Definite Programming

Maximize C·XSubject to

AX=bmatrix(X) is positive semi-

definitewhere X is a vector with size n2, and

matrix(X) is a n by n matrix reshaped from X

Page 6: Nonlinear Dimension Reduction:

Semi-Definite Programming

Constraints: Maintain the distance between neighbors

|Xi-Xj|2=|Yi-Yj|2 for each pair of neighbor (i,j)

Kii+Kjj-Kij-Kji= Gii+Gjj-Gij-Gji where K=YTY,G=XTX

Constrain the output centered on the originΣYi=0 ΣKij=0

K is positive semidefinite

Page 7: Nonlinear Dimension Reduction:

Semi-Definite Programming

Objective function Maximize the sum of pairwise

squared distance between outputsΣij|Yi-Yj|2 Tr(K)

Page 8: Nonlinear Dimension Reduction:

Semi-Definite Programming

Solve the best K using any SDP solver CSDP (fast, stable) SeDuMi (stable, slow) SDPT3 (new, fastest, not well tested)

Page 9: Nonlinear Dimension Reduction:

Locally Linear Embedding

Page 10: Nonlinear Dimension Reduction:

Swiss Roll

N=800

SDE, k=4

LLE, k=18

Page 11: Nonlinear Dimension Reduction:

LLE on Swiss Roll, varying K

K=5 K=6

K=8 K=10

Page 12: Nonlinear Dimension Reduction:

LLE on Swiss Roll, varying K

K=12 K=14

K=16 K=18

Page 13: Nonlinear Dimension Reduction:

LLE on Swiss Roll, varying K

K=20 K=30

K=40 K=60

Page 14: Nonlinear Dimension Reduction:

Twos

N=638

SDE, k=4

LLE, k=18

Page 15: Nonlinear Dimension Reduction:

Teapots

N=400

SDE, k=4

LLE, k=12

Page 16: Nonlinear Dimension Reduction:

LLE on Teapot, varying N

N=400 K=200

K=100 K=50

Page 17: Nonlinear Dimension Reduction:

Faces

N=1900

SDE, failed

LLE, k=12

Page 18: Nonlinear Dimension Reduction:

SDE versus LLE

Similar idea First, compute neighborhoods in the

input space Second, construct a square matrix to

characterize local relationship between input data.

Finally, compute low-dimension embedding using the eigenvectors of the matrix

Page 19: Nonlinear Dimension Reduction:

SDE versus LLE

Different performance SDE: good quality, more robust to

sparse samples, but optimization is slow and hard to scale to large data set

LLE: fast, scalable to large data set, but low quality when samples are sparse, due to locally linear assumption