matrix extensions to sparse recovery yi ma 1,2 allen yang 3 john wright 1 cvpr tutorial, june 20,...

56
Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 200 1 Microsoft Research Asia 3 University of California Berkeley 2 University of Illinois at Urbana- Champaign

Post on 19-Dec-2015

217 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

Matrix Extensions to Sparse Recovery

Yi Ma1,2 Allen Yang3 John Wright1

CVPR Tutorial, June 20, 2009

1Microsoft Research Asia

3University of California Berkeley

2University of Illinois at Urbana-Champaign

Page 2: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

FINAL TOPIC – Generalizations: sparsity to degeneracy

The tools and phenomena underlying sparse recovery generalize very nicely to low-rank matrix recovery

???

Page 3: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

FINAL TOPIC – Generalizations: sparsity to degeneracy

The tools and phenomena underlying sparse recovery generalize very nicely to low-rank matrix recovery

Matrix completion: Given an incomplete subset of the entries of a low-rank matrix, fill in the missing values.

Robust PCA: Given a low-rank matrix which has been grossly corrupted, recover the original matrix.

Page 4: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

???

Face images

Degeneracy: illumination models

Errors: occlusion, corruption

Relevancy data

Degeneracy: user preferences co-predict

Errors: Missing rankings, manipulation

Video

Degeneracy: temporal, dynamic structures

Errors: anomalous events, mismatches…

Examples of degenerate data:

THIS TALK – From sparse recovery to low-rank recovery

Page 5: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

KEY ANALOGY – Connections between rank and sparsity

Sparse recovery Rank minimization

Unknown Vector x Matrix A

Observations y = Ax y = L[A] (linear map)

Combinatorial objective

Convex relaxation

Algorithmic tools

Linear programming Semidefinite programming

Page 6: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

KEY ANALOGY – Connections between rank and sparsity

This talk: exploiting this connection for matrix completion and RPCA

Sparse recovery Rank minimization

Unknown Vector x Matrix A

Observations y = Ax y = L[A] (linear map)

Combinatorial objective

Convex relaxation

Algorithmic tools

Linear programming Semidefinite programming

Page 7: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

CLASSICAL PCA – Fitting degenerate data

If degenerate observations are stacked as columns of a matrix

then

Page 8: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

CLASSICAL PCA – Fitting degenerate data

If degenerate observations are stacked as columns of a matrix

then

Principal Component Analysis via singular value decomposition:

• Stable, efficient computation

• Optimal estimate of under iid Gaussian noise

• Fundamental statistical tool, huge impact in vision, search, bioinformatics

Page 9: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

CLASSICAL PCA – Fitting degenerate data

If degenerate observations are stacked as columns of a matrix

then

But… PCA breaks down under even a single corrupted observation.

Principal Component Analysis via singular value decomposition:

• Stable, efficient computation

• Optimal estimate of under iid Gaussian noise

• Fundamental statistical tool, huge impact in vision, search, bioinformatics

Page 10: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

ROBUST PCA – Problem formulation

……

D - observation A – low-rank E – sparse error

Properties of the errors:

• Each multivariate data sample (column) may be corrupted in some entries

• Corruption can be arbitrarily large in magnitude (not Gaussian!)

Problem: Given recover .A0

Low-rank structure Sparse errors

Page 11: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

ROBUST PCA – Problem formulation

Problem: Given recover .A0

Low-rank structure Sparse errors

Numerous heuristic methods in the literature:

• Random sampling [Fischler and Bolles ‘81] • Multivariate trimming [Gnanadesikan and Kettering ‘72] • Alternating minimization [Ke and Kanade ‘03] • Influence functions [de la Torre and Black ‘03]

No polynomial-time algorithm with strong performance guarantees!

……

D - observation A – low-rank E – sparse error

Page 12: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

ROBUST PCA – Semidefinite programming formulation

Seek the lowest-rank that agrees with the data up to some sparse error:A

Page 13: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

ROBUST PCA – Semidefinite programming formulation

Seek the lowest-rank that agrees with the data up to some sparse error:

Not directly tractable, relax:

A

Page 14: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

ROBUST PCA – Semidefinite programming formulation

Seek the lowest-rank that agrees with the data up to some sparse error:

Not directly tractable, relax:

A

Semidefinite program, solvable in polynomial time

Convex envelope over

Page 15: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

MATRIX COMPLETION – Motivation for the nuclear norm

Related problem: we observe only a small known subset

of entries of a rank- matrix . Can we exactly recover ?

Page 16: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

MATRIX COMPLETION – Motivation for the nuclear norm

Related problem: recover a rank matrix from a known subset of entries

Convex optimization heuristic [Candes and Recht] :

Spectral trimming also succeeds with for

For incoherent , exact recovery with

[Keshavan, Montanari and Oh]

[Candes and Tao]

Page 17: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

ROBUST PCA – Exact recovery?

CONJECTURE: If with sufficiently low-rank and

exactly recovers .

Sparsity of error

sufficiently sparse, then solving

Empirical evidence: probability of correct recovery vs rank and sparsity

Perfect recovery

Rank

Page 18: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

Decompose as or ?

ROBUST PCA – Which matrices and which errors?

Fundamental ambiguity – very sparse matrices are also low-rank:

rank-1 rank-00-sparse 1-sparse

Obviously we can only hope to uniquely recover that are incoherent with the standard basis.

Can we recover almost all low-rank matrices from almost all sparse errors?

Page 19: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

ROBUST PCA – Which matrices and which errors?

Random orthogonal model (of rank r) [Candes & Recht ‘08]:

independent samples from invariant measure on Steifel manifold of orthobases of rank r.

arbitrary.

Page 20: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

ROBUST PCA – Which matrices and which errors?

Random orthogonal model (of rank r) [Candes & Recht ‘08]:

independent samples from invariant measure on Steifel manifold of orthobases of rank r.

arbitrary.

Bernoulli error signs-and-support (with parameter ):

Magnitude of is arbitrary.

Page 21: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

MAIN RESULT – Exact Solution of Robust PCA

“Convex optimization recovers almost any matrix of rank from errors affecting of the observations!”

Page 22: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

BONUS RESULT – Matrix completion in proportional growth

“Convex optimization exactly recovers matrices of rank , even with entries missing!”

Page 23: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

MATRIX COMPLETION – Contrast with literature

• [Candes and Tao 2009]:

Correct completion whp for

Does not apply to the large-rank case

• This work:

Correct completion whp for even with

Proof exploits rich regularity and independence in random orthogonal model.

Caveats:

- [C-T ‘09] tighter for small r. - [C-T ‘09] generalizes better to other matrix ensembles.

Page 24: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

MAIN RESULT – Exact Solution of Robust PCA

“Convex optimization recovers almost any matrix of rank from errors affecting of the observations!”

Page 25: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

ROBUST PCA – Solving the convex program

Semidefinite program in millions of unknowns.

Scalable solution: apply a first-order method with convergence to

Sequence of quadratic approximations [Nesterov, Beck & Teboulle]:

Solved via soft thresholding (E), and singular value thresholding (A).

Page 26: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

ROBUST PCA – Solving the convex program

• Iteration complexity for suboptimal solution.

• Dramatic practical gains from continuation

Page 27: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

SIMULATION – Recovery in various growth scenarios

Correct recovery with and fixed, increasing.

Empirically, almost constant number of iterations:

Provably robust PCA at only a constant factor more computation than conventional PCA.

Page 28: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

SIMULATION – Phase Transition in Rank and Sparsity

Fraction of successes with , varying (10 trials)

Fraction of successes with , varying (65 trials)

[0,.5] x [0,.5][0,1] x [0,1]

[0,1] x [0,1] [0,.4] x [0,.4]

Page 29: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Background modeling from video

Video Low-rank appx. Sparse errorStatic camera surveillance video

200 frames, 72 x 88 pixels,

Significant foregroundmotion

Page 30: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Background modeling from video

Video Low-rank appx. Sparse errorStatic camera surveillance video

550 frames, 64 x 80 pixels,

significant illuminationvariation

Background variation

Anomalous activity

Page 31: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Faces under varying illumination

…RPCA

29 images of one person under varying lighting:

Page 32: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Faces under varying illumination

…RPCA

29 images of one person under varying lighting:

Self- shadowing

Specularity

Page 33: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Initial alignment, inappropriate for recognition:

Page 34: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Page 35: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Page 36: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Page 37: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Page 38: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Page 39: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Page 40: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Page 41: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Page 42: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Page 43: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Final result: per-pixel alignment

Page 44: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

EXAMPLE – Face tracking and alignment

Final result: per-pixel alignment

Page 45: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

SIMULATION – Phase Transition in Rank and Sparsity

Fraction of successes with , varying (10 trials)

Fraction of successes with , varying (65 trials)

[0,.5] x [0,.5][0,1] x [0,1]

[0,1] x [0,1] [0,.4] x [0,.4]

Page 46: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

CONJECTURES – Phase Transition in Rank and Sparsity

1

100

Hypothesized breakdown behavior as m ∞

Page 47: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

CONJECTURES – Phase Transition in Rank and Sparsity

1

100

What we know so far:

This work

Classical PCA

Page 48: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

CONJECTURES – Phase Transition in Rank and Sparsity

1

100

CONJECTURE I: convex programming succeeds in proportional growth

Page 49: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

CONJECTURES – Phase Transition in Rank and Sparsity

1

100

CONJECTURE II: for small ranks ,

any fraction of errors can eventually be corrected.

Similar to Dense Error Correction via L1 Minimization, Wright and Ma ‘08

Page 50: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

CONJECTURES – Phase Transition in Rank and Sparsity

1

100

CONJECTURE III: for any rank fraction, ,

there exists a nonzero fraction of errors that can eventually be

corrected with high probability.

Page 51: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

CONJECTURES – Phase Transition in Rank and Sparsity

1

100

CONJECTURE IV: there is an asymptotically sharp phase transition

between correct recovery with overwhelming probability, and

failure with overwhelming probability.

Page 52: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

CONJECTURES – Connections to Matrix Completion

Our results also suggest the possibility of a proportional growth phase transition for matrix completion.

1

100

• How do the two breakdown points compare?

• How much is gained by knowing the location of the corruption?

Robust PCA

Matrix Completion

Similar to Recht, Xu and Hassibi ‘08

Matrix CompletionRobust PCA

Page 53: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

FUTURE WORK – Stronger results on RPCA?

• RPCA with noise and errors:

Tradeoff between estimation error and robustness to corruption?

• Deterministic conditions on the matrix

• Simultaneous error correction and matrix completion:

bounded noise(e.g., Gaussian)

Conjecture: stable recovery with

we observe

Page 54: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

• Faster algorithms:

Smarter continuation strategies

Parallel implementations, GPU, multi-machine

• Further applications:

Computer vision: photometric stereo, tracking, video repair

Relevancy data: search, ranking and collaborative filtering

Bioinformatics

System Identification

FUTURE WORK – Algorithms and Applications

Page 55: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

• Reference: Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Matrices by Convex Optimization submitted to the Journal of the ACM

• Collaborators: Prof. Yi Ma (UIUC, MSRA)

Dr. Zhouchen Lin (MSRA)

Dr. Shankar Rao (UIUC)

Arvind Ganesh (UIUC)

Yigang Peng (MSRA)

• Funding:

Microsoft Research Fellowship (sponsored by Live Labs)Grants NSF CRS-EHS-0509151, NSF CCF-TF-0514955, ONR YIP N00014-04-1-0633, NSF IIS 07-03756

REFERENCES + ACKNOWLEDGEMENT

Page 56: Matrix Extensions to Sparse Recovery Yi Ma 1,2 Allen Yang 3 John Wright 1 CVPR Tutorial, June 20, 2009 1 Microsoft Research Asia 3 University of California

Questions, please?

THANK YOU!

John Wright Robust PCA: Exact Recovery of Corrupted Low-Rank Matrices