sparse image representation via combined transforms - Convex ...
sparse image representation via combined transforms - Convex ...
sparse image representation via combined transforms - Convex ...
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
4.6. HOMOTOPY 91<br />
A way to interpret the above result is that for the residual r = y − Φˆx, the maximum<br />
amplitude of the analysis transform of the residual ‖Φ T r‖ ∞ is upper bounded by λ 2 . Hence<br />
if λ is small enough and if Φ T is norm preserving—each column of Φ has almost the same l 2<br />
norm—then the de<strong>via</strong>tion of the reconstruction based on ̂x, Φ̂x, from the <strong>image</strong> y is upper<br />
bounded by a small quantity (literally λ 2<br />
) at each direction given by the columns of Φ. So<br />
if we choose a small λ, the corresponding reconstruction cannot be much different from the<br />
original <strong>image</strong>.<br />
After choosing ρ and λ, we have the Hessian and gradient of the objective function (later<br />
denoted by f(x)). For fixed vector x k =(x k 1 ,xk 2 ,... ,xk N )T ∈ R N , the gradient at x k is<br />
⎛<br />
g(x k )=−2Φ T y +2Φ T Φx k + λ ⎜<br />
⎝<br />
¯ρ ′ (x k 1 )<br />
⎞<br />
.<br />
⎟<br />
⎠ , (4.7)<br />
¯ρ ′ (x k N )<br />
where x k i is the i-th element of vector x k , i =1, 2,... ,N, and the Hessian is the matrix<br />
⎛<br />
H(x k )=2Φ T Φ+λ ⎜<br />
⎝<br />
¯ρ ′′ (x k 1 ) ⎞<br />
. .. ⎟<br />
⎠ , (4.8)<br />
¯ρ ′′ (x k N )<br />
where ¯ρ ′′ is the second derivative of function ¯ρ.<br />
4.6 Homotopy<br />
Based on the previous choice of ρ and ¯ρ, when the parameter γ goes to +∞, ρ(x, γ) =<br />
∑ N<br />
i=1 ¯ρ(x i,γ) goes to function ‖x‖ 1 . Note our ultimate goal is to solve the minimum l 1<br />
norm problem (P 1 ). Considering the Lagrangian multiplier method, for a fixed λ, wesolve<br />
minimize ‖y − Φx‖ 2<br />
x<br />
2 + λ‖x‖ 1 . (4.9)<br />
When γ goes to +∞, will the solution of the problem in (4.2) converge to the solution of<br />
the problem in (4.9)?<br />
We prove the convergence under the following two assumptions. The first one is easy to<br />
satisfy. The second one seems too rigorous and may not be true for many situations that we