MA5232 Modeling and Numerical Simulations: Iterative Methods For Mixture-Model Segmentation 8 Apr 2015
MA5232 Modeling and Numerical Simulations: Iterative Methods For Mixture-Model Segmentation 8 Apr 2015
Last time
PCA reduces dimensionality of a data set while
retaining as much as possible the data variation.
Statistical view: The leading PCs are given by the
leading eigenvectors of the covariance.
Geometric view: Fitting a d-dim subspace model via
SVD
Extensions of PCA
Probabilistic PCA via MLE
Kernel PCA via kernel functions and kernel matrices
National University of Singapore
4/7/2015
This lecture
Review basic iterative algorithms for central
clustering
Formulation of the subspace segmentation
problem
Segmentation by Clustering
From: Object Recognition as Machine Translation, Duygulu, Barnard, de Freitas, Forsyth, ECCV02
4/7/2015
Example 4.1
Euclidean distance-based clustering is not
invariant to linear transformation
Central Clustering
Assume data sampled from a mixture of
Gaussian
4/7/2015
4/7/2015
Algorithm
A chicken-and-egg view
Two-Step Iteration
10
4/7/2015
Example
http://util.io/k-means
11
Feature Space
Source: K. Grauman
4/7/2015
Image
Clusters on intensity
Clusters on color
4/7/2015
15
Characteristics of K-Means
It is a greedy algorithm, does not guarantee to
converge to the global optimum.
Given fixed initial clusters/ Gaussian models, the
iterative process is deterministic.
Result may be improved by running k-means
multiple times with different starting conditions.
The segmentation-estimation process can be
treated as a generalized expectationmaximization algorithm
National University of Singapore
16
4/7/2015
17
18
4/7/2015
19
20
10
4/7/2015
Observation:
21
22
11
4/7/2015
23
24
12
4/7/2015
Update
Recall
Assume
is fixed, then maximize the
expected complete log-likelihood
25
26
13
4/7/2015
Exer 4.2
27
EM Algorithm
28
14
4/7/2015
29
w1
30
15
4/7/2015
w1
31
w2
w1
32
16
4/7/2015
w2
w1
33
w2
w1
34
17
4/7/2015
w2
w1
35
Visual example of EM
18
4/7/2015
Potential Problems
Incorrect number of Mixture Components
Singularities
19
4/7/2015
Singularities
A minority of the data can have a
disproportionate effect on the model
likelihood.
For example
20
4/7/2015
GMM example
Singularities
When a mixture component collapses on a
given point, the mean becomes the point, and
the variance goes to zero.
Consider the likelihood function as the
covariance goes to zero.
The likelihood approaches infinity.
21
4/7/2015
K-means VS EM
43
So far
K-means
Expectation Maximization
44
22
4/7/2015
Next up
Multiple-Subspace Segmentation
K-subspaces
EM for Subspaces
45
Multiple-Subspace Segmentation
46
23
4/7/2015
K-subspaces
47
K-subspaces
With noise, we minimize
48
24
4/7/2015
K-subspaces
49
K-subspaces
50
25
4/7/2015
K-subspaces
51
K-subspaces
52
26
4/7/2015
EM for Subspaces
53
EM for Subspaces
54
27
4/7/2015
EM for Subspaces
55
EM for Subspaces
56
28
4/7/2015
EM for Subspaces
57
EM for Subspaces
In the M step
58
29
4/7/2015
EM for Subspaces
59
EM for Subspaces
60
30
4/7/2015
EM for Subspaces
61
EM for Subspaces
62
31
4/7/2015
63
Homework
Read the handout Chapter 4 Iterative
Methods for Multiple-Subspace
Segmentation.
Complete exercise 4.2 (page 111) of the
handout
64
32