manifold blurring mean shift algorithms for...

1
0 2 4 6 8 10 12 14 16 784 2.2 2.4 2.6 2.8 3 3.2 L test set error (%) GBMS baseline 0 50 100 150 200 250 300 350 2.2 2.4 2.6 2.8 3 3.2 k test set error (%) 0 500 1000 2000 3000 4000 2.2 2.4 2.6 2.8 3 3.2 σ test set error (%) baseline LTP 4 6 8 10 20 40 60 0 1 2 3 4 5 6 7 8 original PCA MBMS LTP GBMS test set error (%) training set size (×10 3 ) Left 3 plots : 5–fold cross-validation error (%) curves with a nearest-neighbor classifier on the entire MNIST training dataset (60k points, thus each fold trains on 48k and tests on 12k) using MBMSk; we selected L =9, k = 140, σ = 695 as final values. Right plot : denoising and classification of the MNIST test set (10k points), by training on the entire training set (rightmost value) and also on smaller subsets of it (errorbars over 10 random subsets). Algorithms (L, k, σ ), all using a k -nn graph: MBMSk (9, 140, 695), LTP (9, 140, ), GBMS (0, 140, 600), and PCA (L = 41). 99.29 0.10 0.10 0.00 0.00 0.10 0.31 0.10 0.00 0.00 0.00 99.47 0.26 0.00 0.09 0.09 0.09 0.00 0.00 0.00 0.68 0.58 96.12 0.48 0.10 0.00 0.19 1.55 0.29 0.00 0.00 0.10 0.20 96.04 0.10 1.88 0.00 0.69 0.69 0.30 0.00 0.71 0.00 0.00 96.13 0.00 0.31 0.51 0.10 2.24 0.11 0.11 0.00 1.35 0.22 96.41 0.56 0.11 0.67 0.45 0.42 0.21 0.00 0.00 0.31 0.52 98.54 0.00 0.00 0.00 0.00 1.36 0.58 0.19 0.39 0.00 0.00 96.50 0.00 0.97 0.62 0.10 0.31 1.44 0.51 1.33 0.31 0.41 94.46 0.51 0.20 0.50 0.10 0.59 0.99 0.50 0.10 1.09 0.10 95.84 a0 a1 a2 a3 a4 a5 a6 a7 a8 a9 b0 b1 b2 b3 b4 b5 b6 b7 b8 b9 2.3 2.2 2.1 2 1.9 1.8 1.7 1.6 1.5 1.4 1.3 1.2 1.1 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 99.39 0.10 0.10 0.00 0.00 0.10 0.20 0.10 0.00 0.00 0.00 99.56 0.26 0.09 0.00 0.00 0.00 0.00 0.00 0.09 0.48 0.00 97.97 0.00 0.10 0.00 0.10 1.07 0.29 0.00 0.00 0.00 0.30 97.23 0.10 0.99 0.00 0.40 0.89 0.10 0.00 0.00 0.00 0.00 97.86 0.00 0.31 0.31 0.10 1.43 0.11 0.00 0.00 0.67 0.22 97.76 0.56 0.11 0.22 0.34 0.21 0.21 0.00 0.00 0.10 0.21 99.16 0.00 0.10 0.00 0.00 0.88 0.49 0.00 0.10 0.10 0.00 97.86 0.00 0.58 0.31 0.00 0.31 0.72 0.41 0.72 0.21 0.31 96.61 0.41 0.10 0.20 0.20 0.40 0.99 0.20 0.10 0.89 0.20 96.73 a0 a1 a2 a3 a4 a5 a6 a7 a8 a9 b0 b1 b2 b3 b4 b5 b6 b7 b8 b9 2.3 2.2 2.1 2 1.9 1.8 1.7 1.6 1.5 1.4 1.3 1.2 1.1 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 -0.10 0.00 0.00 0.00 0.00 0.00 0.10 0.00 0.00 0.00 0.00 -0.09 0.00 -0.09 0.09 0.09 0.09 0.00 0.00 -0.09 0.19 0.58 -1.84 0.48 0.00 0.00 0.10 0.48 0.00 0.00 0.00 0.10 -0.10 -1.19 0.00 0.89 0.00 0.30 -0.20 0.20 0.00 0.71 0.00 0.00 -1.73 0.00 0.00 0.20 0.00 0.81 0.00 0.11 0.00 0.67 0.00 -1.35 0.00 0.00 0.45 0.11 0.21 0.00 0.00 0.00 0.21 0.31 -0.63 0.00 -0.10 0.00 0.00 0.49 0.10 0.19 0.29 -0.10 0.00 -1.36 0.00 0.39 0.31 0.10 0.00 0.72 0.10 0.62 0.10 0.10 -2.16 0.10 0.10 0.30 -0.10 0.20 0.00 0.30 0.00 0.20 -0.10 -0.89 a0 a1 a2 a3 a4 a5 a6 a7 a8 a9 b0 b1 b2 b3 b4 b5 b6 b7 b8 b9 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 -0.1 -0.2 -0.3 -0.4 -0.5 -0.6 -0.7 -0.8 -0.9 The confusion matrices (before denoising, after denoising, before minus after). Each row shows the probability (in percentage) of recognizing each digit to all digits. On the left two matrices, white means zero error (perfect classification) and red means positive error. On the right matrix, white means no change (same error before and after denoising), green means denoising reduced the error, and red means denoising increased the error. 0 0 6 3 3 8 4 4 9 5 5 6 7 7 1 9 9 7 1 5 1 2 0 2 2 1 2 3 5 3 3 1 3 4 9 4 5 3 5 6 5 6 6 0 6 7 1 7 8 3 8 8 5 8 9 4 9 9 5 9 Some misclassified images. Each triplet is (test,original-nearest-neighbor,denoised-nearest-neighbor) and the correspond- ing label is above each image, with errors underlined. After denoising there are fewer errors, some of which are arguably wrong ground-truth labels. Classification of MNIST handwritten digits Sample pairs of (original,denoised) images from the training set. The digits look smoother (as if they had been anti-aliased to reduce pixelation) and easier to read, and show sophisti- cated “corrections” (e.g. inpainting and erasure). Comparing the original vs the denoised , one sees this would help classification. τ =0 τ =1 τ =2 τ =3 τ =5 view 0 -10 -5 0 5 10 15 0 10 20 -10 -5 0 5 10 15 -10 -5 0 5 10 15 0 10 20 -10 -5 0 5 10 15 -10 -5 0 5 10 15 0 10 20 -10 -5 0 5 10 15 -10 -5 0 5 10 15 0 10 20 -10 -5 0 5 10 15 -10 -5 0 5 10 15 0 10 20 -10 -5 0 5 10 15 0 10 20 30 40 0 1000 2000 3000 4000 0 1 3 5 7 9 0 0.05 0 500 1000 frequency λ 0 10 20 30 40 0 200 400 600 800 1000 0 1 3 5 7 9 frequency λ 0 2 4 6 0 1000 2000 3000 4000 0 1 3 5 7 9 0 0.05 0 1000 2000 frequency λ view 1 -10 -5 0 5 10 15 -10 -5 0 5 10 15 -2 0 2 -10 -5 0 5 10 15 -10 -5 0 5 10 15 -10 -5 0 5 10 15 -2 0 2 -10 -5 0 5 10 15 -10 -5 0 5 10 15 -10 -5 0 5 10 15 -2 0 2 -10 -5 0 5 10 15 -10 -5 0 5 10 15 -10 -5 0 5 10 15 -2 0 2 -10 -5 0 5 10 15 -10 -5 0 5 10 15 -10 -5 0 5 10 15 -2 0 2 -10 -5 0 5 10 15 Isomap -60 -40 -20 0 20 40 60 -50 -40 -30 -20 -10 0 10 20 30 40 50 -60 -40 -20 0 20 40 60 -50 -40 -30 -20 -10 0 10 20 30 40 50 -60 -40 -20 0 20 40 60 -50 -40 -30 -20 -10 0 10 20 30 40 50 -60 -40 -20 0 20 40 60 -50 -40 -30 -20 -10 0 10 20 30 40 50 -60 -40 -20 0 20 40 60 -50 -40 -30 -20 -10 0 10 20 30 40 50 LTSA -0.03 -0.025 -0.02 -0.015 -0.01 -0.005 0 0.005 0.01 0.015 -0.025 -0.02 -0.015 -0.01 -0.005 0 0.005 0.01 0.015 0.02 0.025 -0.015 -0.01 -0.005 0 0.005 0.01 0.015 0.02 0.025 0.03 -0.025 -0.02 -0.015 -0.01 -0.005 0 0.005 0.01 0.015 0.02 0.025 -0.01 0 0.01 0.02 0.03 -0.02 -0.01 0 0.01 0.02 0.03 -0.02 -0.01 0 0.01 0.02 0.03 -0.03 -0.02 -0.01 0 0.01 0.02 0.03 -0.02 -0.01 0 0.01 0.02 0.03 -0.03 -0.02 -0.01 0 0.01 0.02 0.03 Dimensionality reduction with Isomap and LTSA for different iterations of MBMSk denoising (10–nearest-neighbor graph, L =2, k = 30, σ =5). τ =0 is the original Swiss roll dataset (N = 4 000 points) lifted to 100 dimensions with additive Gaussian noise of stdev 0.6 in each dimension. Isomap/LTSA used a 10-nn graph. Isomap’s residual variances (τ = 0, 1, 2, 3, 5): 0.3128, 0.0030, 0.0002, 0.0002, 0.0003. Right column: histograms over all data points of the normal, tangential, and normal/tangential ratio of the variances (curves for iterations τ =0, 1, 3, 5, 7, 9). Experimental results τ =0 τ =1 τ =2 τ =3 τ =5 τ = 10 τ = 20 τ = 60 MBMSf MBMSk LTP GBMS Denoising a spiral with outliers over iterations (τ =0 is the original dataset). Each box is the square [30, 30] 2 , where 100 outliers were uniformly added to an existing 1 000-point noisy spiral. Algorithms (L, k, σ ): (1, 10, 1.5) and full graph (MBMSf), (1, 10, 1.5) and k -nn graph (MBMSk), (1, 10, ) and k -nn graph (LTP), and (0, ·, 1.5) and full graph (GBMS). 8 Conclusion With adequate parameter values, the proposed MBMS algorithm is very effective at denoising in a handful of itera- tions a dataset with low-dim. structure, even with extreme outliers, and causing very small manifold distortion. It is nonparametric and deterministic (no local optima); its only user parameters (L, k, σ ) are intuitive and good regions for them seem easy to find. LTP (local tangent projection) is a particular, simple case of MBMS that has quasi-optimal performance and only needs L and k . Preprocessing with MBMS improves the quality of algorithms for manifold learning and classification that are sensitive to noise or outliers. We expect MBMS to help in other settings with noisy data of intrinsic low dimensionality, such as density estimation, regression or semi-supervised learning. Work supported by NSF CAREER award IIS–0754089. 9 Computational complexity O(N 2 D + N (D + k ) min(D,k ) 2 ) per iteration, where the first term is for finding nearest neighbors and for the mean-shift step, and the second for the local PCAs. Various accelerations possible, e.g. not updating the k -nn graph affects the result little and makes the cost linear on N . 7 Stopping criteria A practical indicator of whether we have achieved significant denoising while preventing shrinkage is the histogram over all data points of the orthogonal variance λ (the sum of the trailing k L eigenvalues of x n ’s local covariance). Its mean decreases drastically in the first few iterations, while the mean of the histogram of the tangential variance λ stabilizes. 6 Convergence results The MBMS algorithms are covariant under rigid motions. Any dataset that is contained in an L-dim. linear manifold is a fixed point of MBMS (since the tangent space coincides with this manifold and tangential motion is removed). A Gaussian distribution converges cubically to the linear manifold defined by its mean and L leading eigenvec- tors (denoising proceeds independently along each principal axis but motion along the L leading eigenvectors is removed). Essentially, MBMS performs GBMS clustering orthogonal to the principal manifold. 5 How to set the parameters? k : the number of nearest neighbors that estimates the local tangent space. It typically grows sublinearly with N . L: local intrinsic dimension. It could be estimated (e.g. using the correlation dimension) but here we fix it. Note L =0 is GBMS, L = D is no motion. σ : related to the level of local noise outside the manifold. The larger σ , the stronger the denoising and the distortion of the manifold shape. Using a k -nn graph limits the motion to near the k nearest neighbors and allows larger σ . Note σ =0 is no motion, σ = is LTP. 4 Varieties of the MBMS algorithm MBMS (L, k, σ ) with full graph or k -nn graph: given X D×N repeat for n =1,...,N N n ←{1,...,N } (full graph: MBMSf) or k nearest neighbors of x n (k -nn graph: MBMSk) x n ←−x n + m∈N n G σ (x n ,x m ) m ∈N n G σ (x n ,x m ) x m mean-shift step X n k nearest neighbors of x n (μ n , U n ) PCA L (X n ) estimate L-dim tangent space at x n x n (I U n U T n )x n subtract parallel motion end X X + X move points until stop return X LTP (L, k ) with k -nn graph: given X D×N repeat for n =1,...,N X n k nearest neighbors of x n (μ n , U n ) PCA L (X n ) estimate L-dim tangent space at x n x n (I U n U T n )(μ n x n ) project point onto tangent space end X X + X move points until stop return X GBMS (k,σ ) with full or k -nn graph: given X D×N repeat for n =1,...,N N n ←{1,...,N } (full graph) or k nearest neighbors of x n (k -nn graph) x n ←−x n + m∈N n G σ (x n ,x m ) m ∈N n G σ (x n ,x m ) x m mean-shift step end X X + X move points until stop return X 3 The MBMS algorithm Local clustering with Gaussian blurring mean shift (GBMS): the blurring mean-shift update moves datapoints to the kernel average of their neighbors: x n m∈N n G σ (x n , x m ) m ∈N n G σ (x n , x m ) x m . The average is over N n = {1,...,N } (full graph) or the k nearest neighbors of x n (k -nn graph), and G σ (x n , x m ) exp ( 1 2 (x n x m ) 2 ) . This step produces denoising. Local tangent space estimation with PCA: local PCA gives the best linear L-dim. manifold in terms of reconstruction error (i.e., orthogonal projection on the manifold): min μ,U m∈N n x m (UU T (x m μ)+ μ) 2 s.t. U T U = I with U D×L , μ D×1 , whose solution is μ =E N n {x} and U = the leading L eigenvec- tors of cov N n {x}. In general, N n need not equal N n . This step maintains the manifold structure. 2 Abstract We propose a new family of algorithms for denoising data assumed to lie on a low-dimensional manifold. The algorithms are based on the blurring mean-shift update, which moves each data point towards its neighbors, but constrain the motion to be orthogonal to the manifold. The re- sulting algorithms are nonparametric, simple to implement and very effective at removing noise while preserving the curvature of the manifold and limiting shrinkage. They deal well with extreme outliers and with variations of density along the manifold. We apply them as preprocessing for dimensionality reduction; and for nearest-neighbor classification of MNIST digits, with consistent improvements up to 36% over the original data. 1 MANIFOLD BLURRING MEAN SHIFT ALGORITHMS FOR MANIFOLD DENOISING. Weiran Wang and Miguel ´ A. Carreira-Perpi ˜ n ´ an. EECS, School of Engineering, University of California, Merced.

Upload: others

Post on 23-Jul-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: MANIFOLD BLURRING MEAN SHIFT ALGORITHMS FOR …ttic.uchicago.edu/~wwang5/papers/cvpr10a-poster.pdf · 2015-04-22 · 0.11 0.00 0.00 0.67 0.22 97.76 0.56 0.11 0.22 0.34 ... LTP (local

0 2 4 6 8 10 12 14 16 784

2.2

2.4

2.6

2.8

3

3.2

L

test

sete

rror

(%)

GBMS

baseline

0 50 100 150 200 250 300 350

2.2

2.4

2.6

2.8

3

3.2

k

test

sete

rror

(%)

0 500 1000 2000 3000 4000

2.2

2.4

2.6

2.8

3

3.2

σ∞

test

sete

rror

(%)

baseline

LTP

4 6 810 20 40 600

1

2

3

4

5

6

7

8

originalPCAMBMSLTPGBMS

test

sete

rror

(%)

training set size (×103)

Left 3 plots: 5–fold cross-validation error (%) curves with a nearest-neighbor classifier on the entire MNIST training dataset (60k points,thus each fold trains on 48k and tests on 12k) using MBMSk; we selected L = 9, k = 140, σ = 695 as final values. Right plot : denoisingand classification of the MNIST test set (10k points), by training on the entire training set (rightmost value) and also on smaller subsetsof it (errorbars over 10 random subsets). Algorithms (L, k, σ), all using a k-nn graph: MBMSk (9, 140, 695), LTP (9, 140,∞), GBMS(0, 140, 600), and PCA (L = 41).

99.29 0.10 0.10 0.00 0.00 0.10 0.31 0.10 0.00 0.00

0.00 99.47 0.26 0.00 0.09 0.09 0.09 0.00 0.00 0.00

0.68 0.58 96.12 0.48 0.10 0.00 0.19 1.55 0.29 0.00

0.00 0.10 0.20 96.04 0.10 1.88 0.00 0.69 0.69 0.30

0.00 0.71 0.00 0.00 96.13 0.00 0.31 0.51 0.10 2.24

0.11 0.11 0.00 1.35 0.22 96.41 0.56 0.11 0.67 0.45

0.42 0.21 0.00 0.00 0.31 0.52 98.54 0.00 0.00 0.00

0.00 1.36 0.58 0.19 0.39 0.00 0.00 96.50 0.00 0.97

0.62 0.10 0.31 1.44 0.51 1.33 0.31 0.41 94.46 0.51

0.20 0.50 0.10 0.59 0.99 0.50 0.10 1.09 0.10 95.84

a0 a1 a2 a3 a4 a5 a6 a7 a8 a9

b0

b1

b2

b3

b4

b5

b6

b7

b8

b9

2.32.22.1 21.91.81.71.61.51.41.31.21.1 10.90.80.70.60.50.40.30.20.1 0

99.39 0.10 0.10 0.00 0.00 0.10 0.20 0.10 0.00 0.00

0.00 99.56 0.26 0.09 0.00 0.00 0.00 0.00 0.00 0.09

0.48 0.00 97.97 0.00 0.10 0.00 0.10 1.07 0.29 0.00

0.00 0.00 0.30 97.23 0.10 0.99 0.00 0.40 0.89 0.10

0.00 0.00 0.00 0.00 97.86 0.00 0.31 0.31 0.10 1.43

0.11 0.00 0.00 0.67 0.22 97.76 0.56 0.11 0.22 0.34

0.21 0.21 0.00 0.00 0.10 0.21 99.16 0.00 0.10 0.00

0.00 0.88 0.49 0.00 0.10 0.10 0.00 97.86 0.00 0.58

0.31 0.00 0.31 0.72 0.41 0.72 0.21 0.31 96.61 0.41

0.10 0.20 0.20 0.40 0.99 0.20 0.10 0.89 0.20 96.73

a0 a1 a2 a3 a4 a5 a6 a7 a8 a9

b0

b1

b2

b3

b4

b5

b6

b7

b8

b9

2.32.22.1 21.91.81.71.61.51.41.31.21.1 10.90.80.70.60.50.40.30.20.1 0

−0.10 0.00 0.00 0.00 0.00 0.00 0.10 0.00 0.00 0.00

0.00 −0.09 0.00 −0.09 0.09 0.09 0.09 0.00 0.00 −0.09

0.19 0.58 −1.84 0.48 0.00 0.00 0.10 0.48 0.00 0.00

0.00 0.10 −0.10 −1.19 0.00 0.89 0.00 0.30 −0.20 0.20

0.00 0.71 0.00 0.00 −1.73 0.00 0.00 0.20 0.00 0.81

0.00 0.11 0.00 0.67 0.00 −1.35 0.00 0.00 0.45 0.11

0.21 0.00 0.00 0.00 0.21 0.31 −0.63 0.00 −0.10 0.00

0.00 0.49 0.10 0.19 0.29 −0.10 0.00 −1.36 0.00 0.39

0.31 0.10 0.00 0.72 0.10 0.62 0.10 0.10 −2.16 0.10

0.10 0.30 −0.10 0.20 0.00 0.30 0.00 0.20 −0.10 −0.89

a0 a1 a2 a3 a4 a5 a6 a7 a8 a9

b0

b1

b2

b3

b4

b5

b6

b7

b8

b9

0.9

0.8

0.7

0.6

0.5

0.4

0.3

0.2

0.1

0

−0.1

−0.2

−0.3

−0.4

−0.5

−0.6

−0.7

−0.8

−0.9

The confusion matrices (before denoising, after denoising, before minus after). Each row showsthe probability (in percentage) of recognizing each digit to all digits. On the left two matrices, whitemeans zero error (perfect classification) and red means positive error. On the right matrix, whitemeans no change (same error before and after denoising), green means denoising reduced theerror, and red means denoising increased the error.

0 0 6 3 3 8 4 4 9 5 5 6 7 7 1

9 9 7 1 5 1 2 0 2 2 1 2 3 5 3

3 1 3 4 9 4 5 3 5 6 5 6 6 0 6

7 1 7 8 3 8 8 5 8 9 4 9 9 5 9

Some misclassified images. Each triplet is (test,original-nearest-neighbor,denoised-nearest-neighbor) and the correspond-ing label is above each image, with errors underlined. After denoising there are fewer errors, some of which are arguablywrong ground-truth labels.

Classification of MNIST handwritten digits Sample pairs of (original,denoised) images from the training set. Thedigits look smoother (as if they had been anti-aliased to reduce pixelation) and easier to read, and show sophisti-cated “corrections” (e.g. inpainting and erasure). Comparing the original vs the denoised

, one sees this would help classification.

τ = 0 τ = 1 τ = 2 τ = 3 τ = 5

view

0

−10 −5 0 5 10 15010

20

−10

−5

0

5

10

15

−10 −5 0 5 10 15010

20

−10

−5

0

5

10

15

−10 −5 0 5 10 15010

20

−10

−5

0

5

10

15

−10 −5 0 5 10 15010

20

−10

−5

0

5

10

15

−10 −5 0 5 10 15010

20

−10

−5

0

5

10

15

0 10 20 30 400

1000

2000

3000

4000

0135790 0.05

0

500

1000

freq

uenc

y

λ⊥

0 10 20 30 400

200

400

600

800

1000

013579

freq

uenc

y

λ‖

0 2 4 60

1000

2000

3000

4000

0135790 0.05

0

1000

2000

freq

uenc

y

λ⊥/λ‖

view

1

−10 −5 0 5 10 15

−10

−5

0

5

10

15

−202

−10

−5

0

5

10

15

−10 −5 0 5 10 15

−10

−5

0

5

10

15

−202

−10

−5

0

5

10

15

−10 −5 0 5 10 15

−10

−5

0

5

10

15

−202

−10

−5

0

5

10

15

−10 −5 0 5 10 15

−10

−5

0

5

10

15

−202

−10

−5

0

5

10

15

−10 −5 0 5 10 15

−10

−5

0

5

10

15

−202

−10

−5

0

5

10

15

Isom

ap

−60 −40 −20 0 20 40 60−50

−40

−30

−20

−10

0

10

20

30

40

50

−60 −40 −20 0 20 40 60−50

−40

−30

−20

−10

0

10

20

30

40

50

−60 −40 −20 0 20 40 60−50

−40

−30

−20

−10

0

10

20

30

40

50

−60 −40 −20 0 20 40 60−50

−40

−30

−20

−10

0

10

20

30

40

50

−60 −40 −20 0 20 40 60−50

−40

−30

−20

−10

0

10

20

30

40

50

LTS

A

−0.03 −0.025 −0.02 −0.015 −0.01 −0.005 0 0.005 0.01 0.015

−0.025

−0.02

−0.015

−0.01

−0.005

0

0.005

0.01

0.015

0.02

0.025

−0.015 −0.01 −0.005 0 0.005 0.01 0.015 0.02 0.025 0.03

−0.025

−0.02

−0.015

−0.01

−0.005

0

0.005

0.01

0.015

0.02

0.025

−0.01 0 0.01 0.02 0.03

−0.02

−0.01

0

0.01

0.02

0.03

−0.02 −0.01 0 0.01 0.02 0.03

−0.03

−0.02

−0.01

0

0.01

0.02

0.03

−0.02 −0.01 0 0.01 0.02 0.03

−0.03

−0.02

−0.01

0

0.01

0.02

0.03

Dimensionality reduction with Isomap and LTSA for different iterations of MBMSk denoising (10–nearest-neighbor graph,L = 2, k = 30, σ = 5). τ = 0 is the original Swiss roll dataset (N = 4 000 points) lifted to 100 dimensions with additiveGaussian noise of stdev 0.6 in each dimension. Isomap/LTSA used a 10-nn graph. Isomap’s residual variances (τ =0, 1, 2, 3, 5): 0.3128, 0.0030, 0.0002, 0.0002, 0.0003. Right column: histograms over all data points of the normal, tangential,and normal/tangential ratio of the variances (curves for iterations τ = 0, 1, 3, 5, 7, 9).

Experimental resultsτ = 0 τ = 1 τ = 2 τ = 3 τ = 5 τ = 10 τ = 20 τ = 60

MB

MS

fM

BM

Sk

LTP

GB

MS

Denoising a spiral with outliers over iterations (τ = 0 is the original dataset). Each box is the square [−30, 30]2,where 100 outliers were uniformly added to an existing 1 000-point noisy spiral. Algorithms (L, k, σ): (1, 10, 1.5) andfull graph (MBMSf), (1, 10, 1.5) and k-nn graph (MBMSk), (1, 10,∞) and k-nn graph (LTP), and (0, ·, 1.5) and fullgraph (GBMS).

8

ConclusionWith adequate parameter values, the proposed MBMS algorithm is very effective at denoising in a handful of itera-tions a dataset with low-dim. structure, even with extreme outliers, and causing very small manifold distortion. It isnonparametric and deterministic (no local optima); its only user parameters (L, k, σ) are intuitive and good regions forthem seem easy to find. LTP (local tangent projection) is a particular, simple case of MBMS that has quasi-optimalperformance and only needs L and k. Preprocessing with MBMS improves the quality of algorithms for manifoldlearning and classification that are sensitive to noise or outliers. We expect MBMS to help in other settings with noisydata of intrinsic low dimensionality, such as density estimation, regression or semi-supervised learning.

Work supported by NSF CAREER award IIS–0754089.

9

Computational complexityO(N2D+N (D+k) min(D, k)2) per iteration, where the first term is for finding nearest neighbors and for the mean-shiftstep, and the second for the local PCAs. Various accelerations possible, e.g. not updating the k-nn graph affects theresult little and makes the cost linear on N .

7

Stopping criteriaA practical indicator of whether we have achieved significant denoising while preventing shrinkage is the histogramover all data points of the orthogonal variance λ⊥ (the sum of the trailing k − L eigenvalues of xn’s local covariance).Its mean decreases drastically in the first few iterations, while the mean of the histogram of the tangential variance λ‖stabilizes.

6

Convergence results• The MBMS algorithms are covariant under rigid motions.

•Any dataset that is contained in an L-dim. linear manifold is a fixed point of MBMS (since the tangent spacecoincides with this manifold and tangential motion is removed).

•A Gaussian distribution converges cubically to the linear manifold defined by its mean and L leading eigenvec-tors (denoising proceeds independently along each principal axis but motion along the L leading eigenvectors isremoved). Essentially, MBMS performs GBMS clustering orthogonal to the principal manifold.

5

How to set the parameters?• k: the number of nearest neighbors that estimates the local tangent space. It typically grows sublinearly with N .

•L: local intrinsic dimension. It could be estimated (e.g. using the correlation dimension) but here we fix it. NoteL = 0 is GBMS, L = D is no motion.

• σ: related to the level of local noise outside the manifold. The larger σ, the stronger the denoising and the distortionof the manifold shape. Using a k-nn graph limits the motion to near the k nearest neighbors and allows larger σ.Note σ = 0 is no motion, σ =∞ is LTP.

4

Varieties of the MBMS algorithmMBMS (L, k, σ) with full graph or k-nn graph: given XD×Nrepeat

for n = 1, . . . , NNn← {1, . . . , N} (full graph: MBMSf) or

k nearest neighbors of xn (k-nn graph: MBMSk)

∂xn← −xn +∑

m∈Nn

Gσ(xn,xm)∑

m′∈NnGσ(xn,xm′)

xm mean-shift step

Xn← k nearest neighbors of xn

(µn,Un)← PCAL(Xn) estimate L-dim tangent space at xn

∂xn← (I−UnUTn )∂xn subtract parallel motion

endX← X + ∂X move points

until stopreturn X

LTP (L, k) with k-nn graph: given XD×Nrepeat

for n = 1, . . . , NXn← k nearest neighbors of xn

(µn,Un)← PCAL(Xn) estimate L-dim tangent space at xn

∂xn← (I−UnUTn )(µn − xn) project point onto tangent space

endX← X + ∂X move points

until stopreturn X

GBMS (k, σ) with full or k-nn graph: given XD×Nrepeat

for n = 1, . . . , NNn← {1, . . . , N} (full graph) or

k nearest neighbors of xn (k-nn graph)

∂xn← −xn +∑

m∈Nn

Gσ(xn,xm)∑

m′∈NnGσ(xn,xm′)

xm mean-shift step

endX← X + ∂X move points

until stopreturn X

3

The MBMS algorithm• Local clustering with Gaussian blurring mean shift (GBMS): the blurring mean-shift update moves

datapoints to the kernel average of their neighbors:

xn←∑

m∈Nn

Gσ(xn,xm)∑

m′∈NnGσ(xn,xm′)

xm.

The average is over Nn = {1, . . . , N} (full graph) or the k nearest neighbors of xn (k-nn graph),and Gσ(xn,xm) ∝ exp

(

− 12(‖xn − xm‖ /σ)2

)

. This step produces denoising.

• Local tangent space estimation with PCA: local PCA gives the best linear L-dim. manifold interms of reconstruction error (i.e., orthogonal projection on the manifold):

minµ,U

m∈N ′n

∥xm − (UU

T (xm − µ) + µ)∥

2

s.t. UTU = I with UD×L, µD×1, whose solution is µ = EN ′n {x} and U = the leading L eigenvec-

tors of covN ′n {x}. In general, Nn need not equal N ′n. This step maintains the manifold structure.

2AbstractWe propose a new family of algorithms for denoising data assumed to lie on a low-dimensionalmanifold. The algorithms are based on the blurring mean-shift update, which moves each datapoint towards its neighbors, but constrain the motion to be orthogonal to the manifold. The re-sulting algorithms are nonparametric, simple to implement and very effective at removing noisewhile preserving the curvature of the manifold and limiting shrinkage. They deal well with extremeoutliers and with variations of density along the manifold. We apply them as preprocessing fordimensionality reduction; and for nearest-neighbor classification of MNIST digits, with consistentimprovements up to 36% over the original data.

1

MANIFOLD BLURRING MEAN SHIFT ALGORITHMS FOR MANIFOLD DENOISING.Weiran Wang and Miguel A. Carreira-Perpinan.

EECS, School of Engineering, University of California, Merced.