Matrix Completion and Recovery
CS 754 Lecture Notes
1
Matrix Completion in Practice: Scenario 1
2
Matrix Completion in Practice: Scenario 2
3
Matrix Completion in Practice: Scenario 2
http://en.wikipedia.org/wiki/Netflix_Prize
4
Matrix Completion in Practice: Scenario 3
5
Matrix Completion in Practice: Scenario 4
6
�
7
Matrix Completion in Practice: Scenario 5
Lambertian Model
8
The term in the circle(s) is the effective surface area as seen from the light source
Many images of one person under different lighting directions: rank 3 matrix
9
If Lambertian model is satisfied!
M = number of pixels of the form (x,y)
m = number of images, each under a different lighting direction
Assumptions: no self-shadows or specularities
A property of these matrices
10
Scenario 6: Low rank matrices!
11
Column vector
A property of these matrices
12
Low-rank matrices are cool!
13
Low-rank matrices are cool!
14
Theorem 1 (Informal Statement)
15
Cool theorem, but … ☹
16
Theorem 2 (Informal Statement)
17
E. Candes and B. Recht, “Exact matrix completion by convex optimization”, Found. Of Comp. Math., 2009
What is the trace-norm of a matrix?
18
More about trace-norm minimization
19
The devil is in the details
20
Coherence of a subspace
21
Coherence of a basis
22
Formal definition of key assumptions
23
In absolute value
Theorem 2 (Formal Statement)
24
the trace-norm minimizer (in the informal statement of theorem 2)
Candes and Recht , Exact matrix completion via convex optimization, https://statweb.stanford.edu/~candes/papers/MatrixCompletion.pdf
Comments on Theorem 2
25
Comments on Theorem 2
26
Comments on Theorem 2
27
Segway: Coupon Collector’s problem
28
Another version of Theorem 2 = Theorem 2A
29
Candes and Plan, Matrix completion under noise, Proceedings of the IEEE
Matrix Completion under noise
30
Matrix Completion under noise
31
Theorem 3
32
Candes and Plan, Matrix completion under noise, Proceedings of the IEEE
Numerical results: Exact matrix completion
33
34
Candes and Recht , Exact matrix completion via convex optimization, https://statweb.stanford.edu/~candes/papers/MatrixCompletion.pdf
Recall for n x n matrices, dr = 2rn + r2.
35
Candes and Recht , Exact matrix completion via convex optimization, https://statweb.stanford.edu/~candes/papers/MatrixCompletion.pdf
36
Candes and Recht , Exact matrix completion via convex optimization, https://statweb.stanford.edu/~candes/papers/MatrixCompletion.pdf
This is actual a compressive low rank matrix recovery experiment, not a matrix completion experiment. See the section on (Compressive) Low Rank Matrix Recovery later on in these slides.
Numerical results: noisy matrix completion
37
Numerical results: noisy matrix completion
38
Red curve: proportional to upper bound on reconstruction error as per the theorem;
Oracle error in green curve: reconstruction error using least squares assuming that the matrix was of rank r
Numerical results: noisy matrix completion
39
A Minimization Algorithm
40
Singular Value Thresholding (SVT)
41
Ref: Cai et al, A singular value thresholding algorithm for matrix completion, SIAM Journal on Optimization, 2010.
The soft-thresholding procedure obeys the following property (which we state w/o proof).
Properties of SVT (stated w/o proof)
42
Properties of SVT (stated w/o proof)
43
Results
44
Results (Data without noise)
45
Results (Noisy Data)
46
Results on real data
47
Results on real data
48
Results on real data
49
Robust Principal Components Analysis
50
Basic Question
51
Why ask this question? Scenario 1
52
Why ask this question? Scenario 1
53
Why ask this question? Scenario 2
54
Why ask this question? Scenario 2
55
Theorem 1 (Informal Statement)
56
This is a convex optimization problem.
Be careful!
57
Be careful!
58
Candes et al, Robust Principal Component Analysis?
Theorem 1 (Formal Statement)
59
Candes et al, Robust Principal Component Analysis?
Comments on Theorem 1
60
Matrix Recovery with Corrupted as well as missing data
61
Theorem 2 (Formal Statement)
62
Candes et al, Robust Principal Component Analysis?
Comments on Theorem 2
63
Theorem 2.1: Noisy RPCA
where δ is such that
64
Constant term
Zhou et al, Stable principal component pursuit, ISIT 2012
Numerical Results on RPCA
65
Numerical Results on RPCA
66
67
Results
Results on background subtraction
69
Results on background subtraction
70
Results on shadow and specularity removal
72
73
Algorithm for Robust PCA
74
Algorithm for Robust PCA
75
Augmentation term
Lagrangian term
ALM: Some intuition
76
The maximum w.r.t. λ will be ∞ unless the constraint is satisfied. Hence these problems are equivalent.
ALM: Some intuition
77
Due to non-smoothness of the max function, the equivalence has little computational benefit. We smooth it by adding another term that penalizes deviations from a prior estimate of the λ parameters.
Maximization w.r.t. λ is now easy
ALM: Some intuition – inequality constraints
78
Maximization w.r.t. λ is now easy
Algorithm for Robust PCA
79
Lagrange matrix
Update of S using soft-thresholding
Update of L using singular-value soft-thresholding
80
Alternating Minimization Algorithm for Robust PCA
Choice of µ is important. The algorithm in the paper simply sets it to 0.25 n1 n2/ ||M||1. In practice, it should be set from a range of candidate values based on match between ||M-L”-S”||2 and the noise level where L”, S” are estimates of L,S. This would require running the algorithm for many value of µ.
(Compressive) Low Rank Matrix Recovery
81
Low rank matrix recovery
82
Low rank matrix recovery
83
Low rank matrix recovery: restricted isometry property (RIP) for linear maps
for all matrices M of rank at the most r.
84
Low rank matrix recovery: Matrix restricted isometry property (RIP)
85
Theorem 1 for low-rank matrix recovery
86
Recht et al, “Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization”
Theorem 2 for low-rank matrix recovery
87
Recht et al, “Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization”
Theorem 2 for low-rank matrix recovery
88
Theorem 3 for low-rank matrix recovery
89
The best rank-r approximation to M.
C1 and C2 are increasing functions of δr(A) in the domain [0,1/3].
Cai and Zhang, “Sharp RIP bound for sparse signal and low-rank matrix recovery”
Comments on Theorem 3
90
Compressive RPCA: Algorithm and an Application
Primarily based on the paper:
Waters et al, “SpaRCS: Recovering Low-Rank and Sparse Matrices
from Compressive Measurements”, NIPS 2011
91
Problem statement
92
Scenarios
93
Objective function: SpaRCS
94
Free parameters
SpaRCS = sparse and low rank decomposition via compressive sampling
SparCS Algorithm
95
Very simple to implement; but requires tuning of K, r parameters; convergence guarantees not established.
SparCS is similar to CoSamp
96
97
This is for compressive measurement vector u = 𝝓x
Results: Phase transition
98
p = #measurements
K = size of the support of S
r = rank of L
Results: Video CS
99
Follows Rice SPC model, independent compressive measurements on each frame of the matrix M representing the video.
Results: Video CS
100
Follows Rice SPC model, independent compressive measurements on each frame of the matrix M representing the video.
Results: Hyperspectral CS
101
Rice SPC model of CS measurements on every spectral band
Results: Robust matrix completion
102
Note that the support of S is a subset of Ω.
Theorem for Compressive PCP
103
Wright et al, “Compressive Principal Component Pursuit”
Q is obtained from the linear span of different independent N(0,1) matrices with iid entries
This means that the #measurements >= #DOF of (L,S) times log2(m)
Theorem for Compressive PCP
104
Wright et al, “Compressive Principal Component Pursuit”
Notion of μ-incoherent L0:
Definition of Q:
Comments
105
Summary
106