+ All Categories
Home > Documents > Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step...

Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step...

Date post: 16-Mar-2020
Category:
Upload: others
View: 6 times
Download: 0 times
Share this document with a friend
33
Joint Blind Deconvolution and Blind Demixing via Nonconvex Optimization Shuyang Ling Department of Mathematics, UC Davis July 19, 2017 Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 1 / 28
Transcript
Page 1: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Joint Blind Deconvolution and Blind Demixingvia Nonconvex Optimization

Shuyang Ling

Department of Mathematics, UC Davis

July 19, 2017

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 1 / 28

Page 2: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Acknowledgements

Research in collaboration with:

Prof.Xiaodong Li (UC Davis)

Prof.Thomas Strohmer (UC Davis)

Dr.Ke Wei (UC Davis)

This work is sponsored by NSF-DMS and DARPA.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 2 / 28

Page 3: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Outline

(a) Blind deconvolution meets blind demixing: applications in imageprocessing and wireless communication

(b) Mathematical models and convex approach

(c) A nonconvex optimization approach towards joint blind deconvolutionand blind demixing

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 3 / 28

Page 4: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

What is blind deconvolution?

What is blind deconvolution?

Suppose we observe a function y which consists of the convolution of twounknown functions, the blurring function f and the signal of interest g ,plus noise w . How to reconstruct f and g from y?

y = f ∗ g + w .

It is obviously a highly ill-posed bilinear inverse problem...

Much more difficult than ordinary deconvolution...but have importantapplications in various fields.

Solvability? What conditions on f and g make this problem solvable?

How? What algorithms shall we use to recover f and g?

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 4 / 28

Page 5: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Why do we care about blind deconvolution?

Image deblurring

Let f be the blurring kernel and g be the original image, then y = f ∗ g isthe blurred image.Question: how to reconstruct f and g from y

=  

y blurred image

fblurring kernel

goriginal image

 = +

+

wnoise

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 5 / 28

Page 6: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Blind deconvolution meets blind demixing

Suppose there are s users and each of them sends a message x i , which isencoded by C i , to a common receiver. Each encoded message g i = C ix i

is convolved with an unknown impulse response function f i .

User1

User𝑖

User𝑠

𝑔$ = 𝐶$𝑥$: signal

⋮𝑦 = ∑ 𝑓3 ∗ 𝑔3 + 𝑤7

38$𝑓3: channel

𝑓$: channel

𝑓7: channel

𝐸𝑠𝑡𝑖𝑚𝑎𝑡𝑒(𝑓$, 𝑥$)

𝐸𝑠𝑡𝑖𝑚𝑎𝑡𝑒(𝑓3, 𝑥3)

𝐸𝑠𝑡𝑖𝑚𝑎𝑡𝑒(𝑓7, 𝑥7)Decoder

𝑓3 ∗ 𝑔3

𝑓$ ∗ 𝑔$

𝑓7 ∗ 𝑔7

𝑔3 = 𝐶3𝑥3: signal

𝑔7 = 𝐶7𝑥7: signal

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 6 / 28

Page 7: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Blind deconvolution and blind demixing

Consider the model:

y =s∑

i=1

f i ∗ g i + w .

This is even more difficult than blind deconvolution (s = 1), since this is a“mixture” of blind deconvolution problems. It also includes phase retrievalas a special case if s = 1 and g i = f i .

More assumptions

Each impulse response f i has maximum delay spread K (compactsupport):

f i (n) = 0, for n > K , f i =

[hi

0

].

Let g i := C ix i be the signal x i ∈ CN encoded by C i ∈ CL×N withL > N. We also require C i to be mutually incoherent by imposingrandomness.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 7 / 28

Page 8: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Mathematical model

Subspace assumption on the frequency domain

Denote F as the L× L DFT matrix.

Let hi ∈ CK be the first K nonzero entries of f i and B be alow-frequency DFT matrix. There holds, f i = Ff i = Bhi .

Let g i := Aix i where Ai := FC i and x i ∈ CN .

Mathematical model

y =s∑

i=1

diag(Bhi )Aix i + w .

Goal: We want to recover (hi , x i )si=1 from (y ,B,Ai )

si=1.

Remark: The degree of freedom for unknowns: s(K + N); number ofconstraints: L.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 8 / 28

Page 9: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Mathematical model

Subspace assumption on the frequency domain

Denote F as the L× L DFT matrix.

Let hi ∈ CK be the first K nonzero entries of f i and B be alow-frequency DFT matrix. There holds, f i = Ff i = Bhi .

Let g i := Aix i where Ai := FC i and x i ∈ CN .

Mathematical model

y =s∑

i=1

diag(Bhi )Aix i + w .

Goal: We want to recover (hi , x i )si=1 from (y ,B,Ai )

si=1.

Remark: The degree of freedom for unknowns: s(K + N); number ofconstraints: L.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 8 / 28

Page 10: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Naive approach

Nonlinear least squares

We may want to try nonlinear least squares approach:

min(hi ,x i )

∥∥∥∥∥s∑

i=1

diag(Bhi )Aix i − y

∥∥∥∥∥2

︸ ︷︷ ︸F (hi ,x i )

.

The objective function is highly nonconvex and more complicatedthan blind deconvolution (s = 1).

Gradient descent might get stuck at local minima.

No guarantees for recoverability.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 9 / 28

Page 11: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Naive approach

Nonlinear least squares

We may want to try nonlinear least squares approach:

min(hi ,x i )

∥∥∥∥∥s∑

i=1

diag(Bhi )Aix i − y

∥∥∥∥∥2

︸ ︷︷ ︸F (hi ,x i )

.

The objective function is highly nonconvex and more complicatedthan blind deconvolution (s = 1).

Gradient descent might get stuck at local minima.

No guarantees for recoverability.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 9 / 28

Page 12: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Convex relaxation and low-rank matrix recovery

Lifting

Let ai ,l be the l-th column of A∗i and bl be the l-th column of B∗.

yl =s∑

i=1

(Bhi )l · (AIx i )l =s∑

i=1

b∗l hix∗i︸︷︷︸rank-1

ai ,l .

Let X i := hix∗i and define the linear operator Ai : CK×N → CL as,

Ai (Z ) := {b∗l Zai ,l}Ll=1 = {⟨Z ,bla∗i ,l

⟩}Ll=1.

Then, there holds y =∑s

i=1Ai (X i ) + w .

See [Candes-Strohmer-Voroninski 13], [Ahmed-Recht-Romberg, 14].

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 10 / 28

Page 13: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Convex relaxation and low-rank matrix recovery

Rank-s matrix recovery

We rewrite y =∑s

i=1 diag(Bhi )Aix i as

yl =

⟨h1x∗1 0 · · · 00 h2x∗2 · · · 0...

.... . .

...0 0 · · · hsx∗s

︸ ︷︷ ︸

rank-s matrix

,

bla∗1,l 0 · · · 0

0 bla∗2,l · · · 0...

.... . .

...0 0 · · · bla∗s,l

Recover a rank-s block diagonal matrix satisfying convex constraints.

Finding such a rank-s matrix is generally an NP-hard problem.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 11 / 28

Page 14: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Low-rank matrix recovery

Nuclear norm minimization

The ground truth is a rank-s block-diagonal matrix. It is natural torecover the solution via solving

mins∑

i=1

‖Z i‖∗ subject tos∑

i=1

Ai (Z i ) = y

where∑s

i=1 ‖Z i‖∗ is the nuclear norm of blkdiag(Z 1, · · · ,Z s).

Question: Can we recover {hi0x∗i0}si=1 exactly?

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 12 / 28

Page 15: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Convex approach

Theorem

Assume that

Let B ∈ CL×K be a partial DFT matrix with B∗B = IK ;

Each Ai is a Gaussian random matrix.

The SDP relaxation is able to recover {(hi0, x i0)}si=1 exactly withprobability at least 1−O(L−γ). Here the number of measurements Lsatifies

[Ling-Strohmer 15] L ≥ Cγs2(K + µ2hN) log3 L;

[Jung-Krahmer-Stoger 17] L ≥ Cγ(s(K + µ2hN)) log3 L

where µ2h = Lmax1≤i≤s‖Bhi0‖2∞‖hi0‖2

.

We can jointly estimate the channels and signals for s users with onesimple convex program.SDP is able to recover {(hi , x i )}si=1 but it is computationallyexpensive.Can we solve this problem simply with gradient descent which alsocomes with theoretic guarantees?

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 13 / 28

Page 16: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Convex approach

Theorem

Assume that

Let B ∈ CL×K be a partial DFT matrix with B∗B = IK ;

Each Ai is a Gaussian random matrix.

The SDP relaxation is able to recover {(hi0, x i0)}si=1 exactly withprobability at least 1−O(L−γ). Here the number of measurements Lsatifies

[Ling-Strohmer 15] L ≥ Cγs2(K + µ2hN) log3 L;

[Jung-Krahmer-Stoger 17] L ≥ Cγ(s(K + µ2hN)) log3 L

where µ2h = Lmax1≤i≤s‖Bhi0‖2∞‖hi0‖2

.

We can jointly estimate the channels and signals for s users with onesimple convex program.SDP is able to recover {(hi , x i )}si=1 but it is computationallyexpensive.Can we solve this problem simply with gradient descent which alsocomes with theoretic guarantees?

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 13 / 28

Page 17: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

A nonconvex optimization approach?

An increasing list of nonconvex approaches to various problems in machinelearning and signal processing:

Phase retrieval: Candes, Li, Soltanolkotabi, Chen, Wright, Sun, etc...

Matrix completion: Sun, Luo, Montanari, etc...

Various problems: Recht, Wainwright, Constantine, etc...

Two-step philosophy for provable nonconvex optimization

(a) Use spectral method to construct a starting point inside “the basin ofattraction”;

(b) Run gradient descent method.

The key is to build up “the basin of attraction”.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 14 / 28

Page 18: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Building “the basin of attraction”

The basin of attraction relies on the following three observations.

Observation 1: Unboundedness of solution

If the pair (hi0, x i0) is a solution to y =∑s

i=1 diag(Bhi0)Aix i0, thenso is the pair (αihi0, α

−1i x i0) for any αi 6= 0.

Thus the blind deconvolution problem always has infinitely manysolutions of this type. We can recover (hi0, x i0) only up to a scalar.

It is possible that ‖hi‖ � ‖x i‖ (vice versa) while ‖hi‖ · ‖x i‖ is fixed.Hence we define Nd0 to balance ‖hi‖ and ‖x i‖:

Nd0 := {{(hi , x i )}si=1 : ‖hi‖ ≤ 2√di0, ‖x i‖ ≤ 2

√di0}.

where di0 = ‖hi0‖‖x i0‖.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 15 / 28

Page 19: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Building “the basin of attraction”

Observation 2: Incoherence

Our numerical experiments have shown that the algorithm’s performancedepends on how much bl (the rows of B) and hi0 are correlated.

µ2h := max1≤i≤s

L‖Bhi0‖2∞‖hi0‖2

, the smaller µh, the better.

Therefore, we introduce the Nµ to control the incoherence:

Nµ := {{hi}si=1 :√L‖Bhi‖∞ ≤ 4µ

√di0}.

“Incoherence” is not a new idea. In matrix completion, we also require theleft and right singular vectors of the ground truth cannot be too “aligned”with those of measurement matrices {bla∗i ,l}1≤l≤L.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 16 / 28

Page 20: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Building “the basin of attraction”

Observation 3: “Close” to the ground truth

We define Nε to quantify closeness of {(hi , x i )}si=1 to true solution, i.e.,

Nε := {{(hi , x i )}si=1 : ‖hix∗i − hi0x∗i0‖F ≤ εdi0}.

We want to find an initial guess close to {(hi0, x i0)}si=1.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 17 / 28

Page 21: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Building “the basin of attraction”

Based on the three observations above, we define thethree neighborhoods:

The basin of attraction

The basin of attraction is the intersection of the following three setsNd0 ∩Nµ ∩Nε:

Nd0 := {{(hi , x i )}si=1 : ‖hi‖ ≤ 2√

di0, ‖x i‖ ≤ 2√di0, 1 ≤ i ≤ s}

Nµ := { {hi}si=1 :√L‖Bhi‖∞ ≤ 4

√di0µ, 1 ≤ i ≤ s}

Nε :=

{{(hi , x i )}si=1 :

‖hix∗i − hi0x∗i0‖Fdi0

≤ ε, 1 ≤ i ≤ s

}where di0 = ‖hi0‖‖x i0‖, µ is a parameter and µ ≥ µh and ε is apredetermined parameter in (0, 1

15 ].

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 18 / 28

Page 22: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Objective function: a variant of projected gradient descent

The objective function F consists of two parts: F and G :

min(h,x)

F (h, x) := F (h, x)︸ ︷︷ ︸least squares term

+ G (h, x)︸ ︷︷ ︸regularization term

where F (h, x) := ‖∑s

i=1Ai (hix∗i )− y‖2 and

G (h, x) := ρ

s∑i=1

[G0

(‖hi‖2

2di

)+ G0

(‖x i‖2

2di

)︸ ︷︷ ︸Nd0

: balance ‖hi‖ and ‖x i‖

+L∑

l=1

G0

(L|b∗l hi |2

8diµ2

)︸ ︷︷ ︸Nµ: impose incoherence

].

Here G0(z) = max{z − 1, 0}2, ρ ≈ d2, d ≈ d0, di ≈ di0 and µ ≥ µh.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 19 / 28

Page 23: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Algorithm: Initialization via spectral method

Note that

A∗i (y) = A∗i Ai (hi0x∗i0)︸ ︷︷ ︸E(A∗i Ai (hi0x∗i0))=hi0x∗i0

+A∗i

∑j 6=i

Aj(hj0x∗j0)

︸ ︷︷ ︸

with mean 0

The leading singular vectors of A∗i (y) can approximate (hi0, x i0).

Step 1: Initialization via spectral method and projection:

1: for i = 1, 2, . . . , s do2: Compute A∗i (y), (since E(A∗i (y)) = hi0x∗i0);

3: (d , hi0, x i0) = svds(A∗i (y));

4: u(0)i := PNµ(

√di hi0) and v (0)

i :=√di x i0.

5: end for

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 20 / 28

Page 24: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Algorithm: Wirtinger gradient descent

Step 2: Gradient descent with constant stepsize η:

1: Initialization: obtain (u(0)i , v (0)

i ) via Algorithm 1.2: for t = 1, 2, . . . , do3: for i = 1, 2, . . . , s do

4: u(t)i = u(t−1)

i − η∇Fhi(u(t−1), v (t−1))

5: v (t)i = v (t−1)

i − η∇Fx i (u(t−1), v (t−1))

6: end for7: end for

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 21 / 28

Page 25: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Main results

Theorem [Ling-Strohmer 17]

Assume w ∼ CN (0, σ2d20/L) and Ai as a complex Gaussian matrix.

There hold:

the initial guess (u(0), v (0)) ∈ 1√3Nd0

⋂ 1√3Nµ⋂N 2ε

5√sκ,√∑s

i=1 ‖u(t)i (v (t)

i )∗ − hi0x∗i0‖2F ≤ (1− α)tεd0 + c0√s‖A∗(w)‖

with probability at least 1− L−γ+1 and α = O((s(K + N) log2 L)−1) if

L ≥ Cγ(µ2h + σ2)s2κ4(K + N) log2 L log s/ε2,

where κ = max di0min di0

.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 22 / 28

Page 26: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Main results

Theorem [Ling-Strohmer 17]

Assume w ∼ CN (0, σ2d20/L) and Ai as a complex Gaussian matrix.

There hold:

the initial guess (u(0), v (0)) ∈ 1√3Nd0

⋂ 1√3Nµ⋂N 2ε

5√sκ,√∑s

i=1 ‖u(t)i (v (t)

i )∗ − hi0x∗i0‖2F ≤ (1− α)tεd0︸ ︷︷ ︸linear convergence

+ c0√s‖A∗(w)‖︸ ︷︷ ︸error term

with probability at least 1− L−γ+1 and α = O((s(K + N) log2 L)−1) if

L ≥ Cγ(µ2h + σ2)s2κ4(K + N) log2 L log s/ε2,

where κ = max di0min di0

.

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 23 / 28

Page 27: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Remark

The iterates (u(t)i , v (t)

i ) converges linearly to (hi0, x i0):

‖u(∞)i (v (∞)

i )∗ − hi0x∗i0‖F ≤ c0√s‖A∗(w)‖

‖A∗(w)‖ converges to 0 with the rate of O(L−1/2):

‖A∗(w)‖ ≤ C0σd0

√s(K + N)(log2 L)

L

Therefore, (u(∞)i , v (∞)

i ) is a consistent estimator of (hi0, x i0).

Challenges: s2 is not optimal. The optimal scaling should beL = O(s(K + N)) instead of L = O(s2(K + N)).

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 24 / 28

Page 28: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Remark

The iterates (u(t)i , v (t)

i ) converges linearly to (hi0, x i0):

‖u(∞)i (v (∞)

i )∗ − hi0x∗i0‖F ≤ c0√s‖A∗(w)‖

‖A∗(w)‖ converges to 0 with the rate of O(L−1/2):

‖A∗(w)‖ ≤ C0σd0

√s(K + N)(log2 L)

L

Therefore, (u(∞)i , v (∞)

i ) is a consistent estimator of (hi0, x i0).

Challenges: s2 is not optimal. The optimal scaling should beL = O(s(K + N)) instead of L = O(s2(K + N)).

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 24 / 28

Page 29: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Numerics: Does L scale linearly with s?

Let each Ai be a complex Gaussian matrix. The number of measurementscales linearly with the number of sources s if K and N are fixed.Approximately, L ≈ 1.5s(K + N) yields exact recovery.

s: 1 to 8

Num

ber

of m

easure

ments

: L, fr

om

100 to 1

250

L vs. s, K = N = 50, Regularized GD, Gaussian

1 2 3 4 5 6 7 8

250

500

750

1000

1250

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Figure: Black: failure; white: success

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 25 / 28

Page 30: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Back to the communication example

A more practical and useful choice of encoding matrix C i : C i = D iH (i.e.,Ai = FD iH) where D i is a diagonal random binary ±1 matrix and H isan L× N deterministic partial Hadamard matrix. With this setting, ourapproach can demix many users without performing channel estimation.

1 2 3 4 5 6 7 8 9 10 11 120

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

s: number of sources

Em

piric

al pro

babili

ty o

f success

Plot: K = N = 50; Hadamard matrix

L = 512

L = 786

L = 1024

L = 1280

L ≈ 1.5s(K + N) yields exact recovery.Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 26 / 28

Page 31: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Numerics: robustness

We see that the relative error is linearly correlated with the noise in dB.Approximately, 10 units of increase in SNR leads to the same amount ofdecrease in relative error (in dB).

0 5 10 15 20 25 30 35 40 45 50−60

−50

−40

−30

−20

−10

0

10

SNR(dB)

Ave

rag

e R

ela

tive

Err

or

of

10

Sa

mp

les(d

B)

K = N = 64, s = 6, Gaussian

L = 2s(K+N)

L = 4s(K+N)

0 5 10 15 20 25 30 35 40 45 50−60

−50

−40

−30

−20

−10

0

10

SNR(dB)

Ave

rag

e R

ela

tive

Err

or

of

10

Sa

mp

les(d

B)

K = N = 64, s = 6, Hadamard

L = 2s(K+N)

L = 4s(K+N)

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 27 / 28

Page 32: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Outlook and Conclusion

Conclusion: The proposed algorithm is arguably the first blinddeconvolution/blind demixing algorithm that is numerically efficient,robust against noise and comes with rigorous recovery guarantees undersubspace conditions.

Open problem: Does similar result hold for other types of Ai?

Open problem: what if either hi or x i is sparse?

Major open problem in nonconvex optimization:

How to remove the s2-dependence for rank-s matrix recovery?

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 28 / 28

Page 33: Joint Blind Deconvolution and Blind Demixing via Nonconvex ...sling/Slides/2017FOCM.pdf · Two-step philosophy for provable nonconvex optimization (a)Use spectral method to construct

Outlook and Conclusion

Conclusion: The proposed algorithm is arguably the first blinddeconvolution/blind demixing algorithm that is numerically efficient,robust against noise and comes with rigorous recovery guarantees undersubspace conditions.

Open problem: Does similar result hold for other types of Ai?

Open problem: what if either hi or x i is sparse?

Major open problem in nonconvex optimization:

How to remove the s2-dependence for rank-s matrix recovery?

Shuyang Ling (UC Davis) FOCM, Barcelona, 2017 July 19, 2017 28 / 28


Recommended