sparse image representation via combined transforms - Convex ...
sparse image representation via combined transforms - Convex ...
sparse image representation via combined transforms - Convex ...
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
72 CHAPTER 3. IMAGE TRANSFORMS AND IMAGE FEATURES<br />
The idea of the BOB is the following: if the entropy associated with the two subintervals is<br />
lower than the entropy associated with the larger dyadic interval, then we choose to divide<br />
the larger dyadic interval. This process is repeated until there is no more partitioning to<br />
do. The final partition of the unit interval [0, 1) determines a subset of localized cosine<br />
functions. It’s not hard to observe that this subset of local cosine functions makes an<br />
orthonormal basis. For a given signal, instead of computing the entropy of a statistical<br />
distribution, we can compute the empirical entropy, which is just the entropy of a set of<br />
coefficients. Note that the coefficients are from a transform of the signal. Based on this<br />
empirical entropy, we can use the BOB algorithm to choose a subset of coefficients, and<br />
correspondingly a subset of basis functions that form a basis.<br />
Complexity. Suppose the length of the signal is N, forN ∈ N. Note that when we take<br />
cosine functions as an orthonormal basis over an interval, the “folding” transform gives<br />
another orthonormal system whose elements are functions that are defined on the entire<br />
real axis, but are localized (have finite support). For every dyadic interval, we have such<br />
an orthonormal system. Combining all these systems, we actually get the cosine packets<br />
(CP). Suppose dyadic intervals having the same length is a cover of the entire real axis. We<br />
say that CP elements associated with these intervals are at the same scale. For length-N<br />
signals, we have log 2 N scales. To calculate the CP coefficients at one scale, we can first<br />
apply “folding” to the data, then apply a DCT. The folding is an O(N) operation. The<br />
discrete cosine transform has O(N log N) complexity. Since we have log 2 N scales, the total<br />
complexity of computing CP coefficients is O(N log 2 N).<br />
Wavelet Packets<br />
In MRA, instead of dividing the low-frequency part, or a scaling space (a space spanned<br />
by scaling functions), we can apply a quadratic mirror filter to divide the high-frequency<br />
part, or a wavelet space (which is spanned by the wavelet functions). This idea unveils<br />
developments of wavelet packets. Equivalently, in the filter banks algorithm, at each step,<br />
instead of deploying a pair of filters after an LPF, we can deploy a pair of filters after an<br />
HPF. By doing this, we partition the high-frequency part. Since a wavelet space contains<br />
high-frequency components, partition in a high-frequency part is equivalent to partition in<br />
a wavelet space.<br />
In every step, we can adaptively choose to deploy a pair of filters either after an LPF<br />
or an HPF. In this framework, all possible sets of coefficients have a binary tree structure.