sparse image representation via combined transforms - Convex ...
sparse image representation via combined transforms - Convex ... sparse image representation via combined transforms - Convex ...
106 CHAPTER 5. ITERATIVE METHODS methods are called Krylov subspace approximations [78, Page 49]. [C2] At step k, the CG error vector and the MINRES residual vector can be written as e k = p C k (A)e 0, r k = p M k (A)r 0, where p C k and pM k are two polynomials with degree no higher than k that take value 1 at the origin. Moreover, ‖e k ‖ A = min‖p k (A)e 0 ‖ A , p k ‖r k ‖ = min‖p k (A)r 0 ‖, p k where p k is a kth-degree polynomial with value 1 at the origin. [C3] Suppose for a Hermitian and positive semidefinite matrix A, A = UΛU T is its eigendecomposition, where U is an orthogonal matrix and Λ is a diagonal matrix. Suppose Λ=diag{λ 1 ,... ,λ N }. We have sharp bounds for the norms of the error and the residual in CG and MINRES: ‖e k ‖ A /‖e 0 ‖ A ≤ min max k(λ i )|, p k i=1,2,... ,N for CG; (5.3) ‖r k ‖/‖r 0 ‖≤min max k(λ i )|, p k i=1,2,... ,N for MINRES. (5.4) In (5.3) and (5.4), if the eigenvalues are tightly clustered around a single point (away from the origin), then the right-hand sides are more likely to be minimized; hence, iterative methods tends to converge quickly. On the other hand, if the eigenvalues are widely spread, especially if they lie on the both sides of the origin, then the values on the right-hand sides of both inequalities are difficult to be minimized; hence, iterative methods may converge slowly. Note that the above results are based on exact arithmetic. In finite-precision computation, these error bounds are in general not true, because the round-off errors that are due to finite precision may destroy assumed properties in the methods (e.g., orthogonality). For further discussion, we refer to Chapter 4 of [78] and the references therein.
5.1. OVERVIEW 107 5.1.4 Preconditioner Before we move into detailed discussion, we would like to point out that the content of this section applies to both CG and MINRES. When the original matrix A does not have a good eigenvalue distribution, a preconditioner may help. In our case, we only need to consider the problem as in (5.2) for Hermitian matrices. We consider solving the preconditioned system L −1 AL −H y = L −1 b, x = L −H y, (5.5) where L is a preconditioner. When the matrix-vector multiplications associated with L −H and L −1 have fast algorithms and L −1 AL −H has a good eigenvalue distribution, the system in (5.2) can be solved in fewer iterations. Before giving a detailed discussion about preconditioners, we need to restate our setting. Here A is the Hessian at step k: A = H(x k ). To simplify the discussion, we assume that Φ is a concatenation of several orthogonal matrices: Φ=[T 1 ,T 2 ,... ,T m ], where T 1 ,T 2 ,... ,T m are n × n orthogonal square matrices. The number of columns of Φ is equal to N = mn 2 . Hence from (4.8), ⎡ ⎤ I + D 1 T1 T T 2 ··· T1 T T m 1 2 A = T2 T T 1 I + D 2 ⎢ . , (5.6) ⎣ . .. . ⎥ ⎦ TmT T 1 ··· I + D m where ⎛ D i = 1 2 λ ⎜ ⎝ ¯ρ ′′ (x 1+(i−1)n2 k ) . .. ¯ρ ′′ (x n2 +(i−1)n 2 k ) ⎞ ⎟ ⎠ , i =1, 2,... ,m, are diagonal matrices. To simplify (5.6), we consider A left and right multiplied by a block
- Page 83 and 84: 3.2. WAVELETS AND POINT SINGULARITI
- Page 85 and 86: 3.2. WAVELETS AND POINT SINGULARITI
- Page 87 and 88: 3.2. WAVELETS AND POINT SINGULARITI
- Page 89 and 90: 3.2. WAVELETS AND POINT SINGULARITI
- Page 91 and 92: 3.2. WAVELETS AND POINT SINGULARITI
- Page 93 and 94: 3.3. EDGELETS AND LINEAR SINGULARIT
- Page 95 and 96: 3.4. OTHER TRANSFORMS 67 uncertaint
- Page 97 and 98: 3.4. OTHER TRANSFORMS 69 Chirplets
- Page 99 and 100: 3.4. OTHER TRANSFORMS 71 Folding. A
- Page 101 and 102: 3.4. OTHER TRANSFORMS 73 We can app
- Page 103 and 104: 3.5. DISCUSSION 75 give only a few
- Page 105 and 106: 3.7. PROOFS 77 the ijth component o
- Page 107 and 108: 3.7. PROOFS 79 Similarly, we have [
- Page 109 and 110: Chapter 4 Combined Image Representa
- Page 111 and 112: 4.2. SPARSE DECOMPOSITION 83 interi
- Page 113 and 114: 4.3. MINIMUM l 1 NORM SOLUTION 85 l
- Page 115 and 116: 4.4. LAGRANGE MULTIPLIERS 87 ρ( x
- Page 117 and 118: 4.5. HOW TO CHOOSE ρ AND λ 89 3 (
- Page 119 and 120: 4.6. HOMOTOPY 91 A way to interpret
- Page 121 and 122: 4.7. NEWTON DIRECTION 93 4.7 Newton
- Page 123 and 124: 4.9. ITERATIVE METHODS 95 1. Avoidi
- Page 125 and 126: 4.11. DISCUSSION 97 ρ(β) =‖β
- Page 127 and 128: 4.12. PROOFS 99 4.12.2 Proof of The
- Page 129 and 130: 4.12. PROOFS 101 case of (4.16). Co
- Page 131 and 132: Chapter 5 Iterative Methods This ch
- Page 133: 5.1. OVERVIEW 105 the k-th iteratio
- Page 137 and 138: 5.2. LSQR 109 among all the block d
- Page 139 and 140: 5.2. LSQR 111 5.2.3 Algorithm LSQR
- Page 141 and 142: 5.3. MINRES 113 2. For k =1, 2,...,
- Page 143 and 144: 5.3. MINRES 115 using the precondit
- Page 145 and 146: 5.4. DISCUSSION 117 From (I + S 1 )
- Page 147 and 148: Chapter 6 Simulations Section 6.1 d
- Page 149 and 150: 6.3. DECOMPOSITION 121 10 5 5 5 20
- Page 151 and 152: 6.4. DECAY OF COEFFICIENTS 123 10 2
- Page 153 and 154: 6.5. COMPARISON WITH MATCHING PURSU
- Page 155 and 156: 6.6. SUMMARY OF COMPUTATIONAL EXPER
- Page 157 and 158: Chapter 7 Future Work In the future
- Page 159 and 160: 7.2. MODIFYING EDGELET DICTIONARY 1
- Page 161 and 162: 7.3. ACCELERATING THE ITERATIVE ALG
- Page 163 and 164: Appendix A Direct Edgelet Transform
- Page 165 and 166: A.2. EXAMPLES 137 edgelet transform
- Page 167 and 168: A.3. DETAILS 139 (a) Stick image (b
- Page 169 and 170: A.3. DETAILS 141 (a) Lenna image (b
- Page 171 and 172: A.3. DETAILS 143 Ordering of Dyadic
- Page 173 and 174: A.3. DETAILS 145 (1,K +1), (1,K +2)
- Page 175 and 176: A.3. DETAILS 147 x 1 , y 1 x 2 , y
- Page 177 and 178: Appendix B Fast Edgelet-like Transf
- Page 179 and 180: B.1. TRANSFORMS FOR 2-D CONTINUOUS
- Page 181 and 182: B.2. DISCRETE ALGORITHM 153 B.2.1 S
- Page 183 and 184: B.2. DISCRETE ALGORITHM 155 extensi
106 CHAPTER 5. ITERATIVE METHODS<br />
methods are called Krylov subspace approximations [78, Page 49].<br />
[C2] At step k, the CG error vector and the MINRES residual vector can be written as<br />
e k = p C k (A)e 0,<br />
r k = p M k (A)r 0,<br />
where p C k and pM k<br />
are two polynomials with degree no higher than k that take value<br />
1 at the origin. Moreover,<br />
‖e k ‖ A = min‖p k (A)e 0 ‖ A ,<br />
p k<br />
‖r k ‖ = min‖p k (A)r 0 ‖,<br />
p k<br />
where p k is a kth-degree polynomial with value 1 at the origin.<br />
[C3] Suppose for a Hermitian and positive semidefinite matrix A, A = UΛU T is its eigendecomposition,<br />
where U is an orthogonal matrix and Λ is a diagonal matrix. Suppose<br />
Λ=diag{λ 1 ,... ,λ N }. We have sharp bounds for the norms of the error and the<br />
residual in CG and MINRES:<br />
‖e k ‖ A /‖e 0 ‖ A ≤ min max k(λ i )|,<br />
p k i=1,2,... ,N<br />
for CG; (5.3)<br />
‖r k ‖/‖r 0 ‖≤min max k(λ i )|,<br />
p k i=1,2,... ,N<br />
for MINRES. (5.4)<br />
In (5.3) and (5.4), if the eigenvalues are tightly clustered around a single point (away<br />
from the origin), then the right-hand sides are more likely to be minimized; hence, iterative<br />
methods tends to converge quickly. On the other hand, if the eigenvalues are widely spread,<br />
especially if they lie on the both sides of the origin, then the values on the right-hand sides<br />
of both inequalities are difficult to be minimized; hence, iterative methods may converge<br />
slowly.<br />
Note that the above results are based on exact arithmetic. In finite-precision computation,<br />
these error bounds are in general not true, because the round-off errors that are due<br />
to finite precision may destroy assumed properties in the methods (e.g., orthogonality). For<br />
further discussion, we refer to Chapter 4 of [78] and the references therein.