Professional Documents
Culture Documents
10 Lowrank Quadratic PDF
10 Lowrank Quadratic PDF
Justin Romberg
Tsinghua University
October 18, 2013
Beijing, China
Matrix completion
R*
X = L
Recommender Euclidean
Systems Embedding
X1,1 − X1,3 − X1,5
− X2,2 − X2,4 −
X = − X3,2 X3,3 − −
X4,1 − − X4,4 X4,5
− − − X5,4 X5,5
X1,1 − X1,3 − X1,5
− X2,2 − X2,4 −
X = − X3,2 X3,3 − −
X4,1 − − X4,4 X4,5
− − − X5,4 X5,5
X1,1 − X1,3 − X1,5
− X2,2 − X2,4 −
X=
− X3,2 X3,3 − −
X4,1 − − X4,4 X4,5
− − − X5,4 X5,5
This particular result is due to Recht ’09, but there are related results from
Candes, Tao, Keshevan, Montenari, Oh, Plan, and others ...
Suppose that an K × N matrix X0 = UΣV∗ is rank R with coherence
K N KN
µ = max max kU∗ ei k22 , max kV∗ ei k22 , kUV∗ k2∞
R i R i R
• Any subset
X= that misses
component
nothing!
• Still need to
X= entire first r
• Want each e
(courtesy of Benjamin Recht)
provide nea
General matrix recovery
We can also consider the more general problem of recovering a matrix
from a set of linear measurements.
A single linear measurement of a K × N matrix X0 can be written as the
trace inner product between X0 and another K × N matrix Am :
ym = hX0 , Am iF = trace(A∗m X0 )
ym = h iF
,
General matrix recovery
ym = hX0 , Am iF = trace(A∗m X0 )
y = A(X0 )
• 2x2 matrices
• plotted in 3d
rank 1
x2 + z2 + 2y2 = 1
Convex hull:
Work by Recht, Fazel, Parrilo, and Mohan has shown that if A obeys the
matrix RIP; for a δ < 1
Work by Recht, Fazel, Parrilo, and Mohan has shown that if A obeys the
matrix RIP; for a δ < 1
The recovery is stable in the presence of noise and is robust for matrices
that are approximately low rank.
Random linear measurements
Just as in the “sparse RIP” case you can couple that with standard bounds
on the entropy of the space of low rank matrices to see that δ2R < 1 when
M & R(K + N )
Recht, Fazel, and Parillo had an early result of this nature in 2007; it was
later refined by Candes and Plan in 2010.
Bilinear equations
2 3
v12 v1 v2 v1 v3 ··· v1 vN
6 v2 v1 v22 v2 v3 ··· v2 vN 7
6 7
T 6 v3 v1 v3 v2 v33 ··· v3 vN 7
vv = 6 7
6 .. .. 7
4 ··· . . 5
2
vN v1 vN v2 vN v3 ··· vN
L ∼ Const · N
for random a` .
A stylized communications problem:
unknown multipath channel
Multipath
m encode p y decode m̃
y = HCm h̃
= m
= discovers unknown
p channel and message
C
y H p
convolution with
unknown channel
Channel coding: multipath
y = HCm
= h(0)(Cm)↓0 + h(1)(Cm)↓1 + · · · + h(K − 1)(Cm)↓K−1
Bad:
With H unknown, these are nonlinear measurements
Good:
There are two sources of structure we can exploit
I the channel h is “short”
I the coded message Cm lives in a known subspace
Channel coding: multipath
y = G1 G2 · · · Gp m(1)h
m(2)h
..
shifts of column n .
m(N )h
Channel coding: multipath
y = G1 G2 · · · Gp m(1)h
m(2)h
..
shifts of column n .
m(N )h
µ2h
ĥ(ω)
0
freq !
N = message length
K = channel length
L = code length
µ2h = channel coherence
A=
G1 G2 · · · Gp