20.05.2014 Views

View - ResearchGate

View - ResearchGate

View - ResearchGate

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

A wavelet approach to cardiac signal<br />

processing for low-power<br />

hardware applications<br />

Joël M.H. Karel


©Copyright Joël M.H. Karel, Maastricht 2009<br />

Universitaire Pers Maastricht<br />

ISBN 978-90-5278-887-6<br />

All rights reserved. No part of this thesis may be reproduced, stored in a retrieval system of<br />

any nature, or transmitted in any form by any means, electronic, mechanical, photocopying,<br />

recording or otherwise, included a complete or partial transcription, without the permission<br />

of the author.


A WAVELET APPROACH TO CARDIAC SIGNAL PROCESSING FOR<br />

A WAVELETLOW-POWER APPROACH TO HARDWARE CARDIAC APPLICATIONS<br />

SIGNAL PROCESSING FOR<br />

LOW-POWER HARDWARE APPLICATIONS<br />

PROEFSCHRIFT<br />

PROEFSCHRIFT<br />

ter verkrijging van de graad van doctor aan de Universiteit Maastricht, op<br />

ter gezag verkrijging van de van Rector de graad Magnificus, van doctor Prof. aan mr. deG.P.M.F. Universiteit Mols Maastricht, volgens het op<br />

besluit gezag van de hetRector CollegeMagnificus, van Decanen, Prof. inmr. hetG.P.M.F. openbaarMols te verdedigen volgens het op<br />

besluit van hetdinsdag College15 vandecember Decanen, 2009 in het omopenbaar 14.00 uur te verdedigen op<br />

dinsdag 15 december 2009 om 14.00 uur<br />

door<br />

door<br />

Joël Matheus Hendrikus Karel<br />

Joël Matheus Hendrikus Karel<br />

UUNIVERSITAIRE<br />

PERS MAASTRICHT<br />

P<br />

M


Promotor:<br />

Prof. dr. ir. R.L.M. Peeters<br />

Copromotor:<br />

Dr. R.L. Westra<br />

Beoordelingscommissie:<br />

Prof. dr. H. Kingma (voorzitter)<br />

Prof. dr. M.P.F. Berger<br />

Prof. dr. Y. Rudy (Washington University in St. Louis)<br />

Dr. ir. W.A. Serdijn (Technische Universiteit Delft)<br />

Dr. P.G.A. Volders<br />

The research reported in this thesis has been funded by Technology Foundation STW (project<br />

number DTC 6418)


Contents<br />

Contents<br />

i<br />

1 Introduction 3<br />

1.1 Research setting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3<br />

1.2 Research questions and thesis outline . . . . . . . . . . . . . . . . . . . . . 4<br />

2 Cardiac signal processing 7<br />

2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7<br />

2.2 The human heart . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8<br />

2.3 Electrocardiogram data . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10<br />

2.3.1 The Electrocardiogram . . . . . . . . . . . . . . . . . . . . . . . . . 10<br />

2.3.2 Signal archives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14<br />

2.4 Cardiac rhythms and pathologies . . . . . . . . . . . . . . . . . . . . . . . 14<br />

2.5 Signal processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15<br />

2.6 Fourier transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15<br />

2.7 Laplace and z-transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . 18<br />

2.8 Linear systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19<br />

2.9 Filtering signals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22<br />

3 Wavelet transformations 25<br />

3.1 Continuous wavelets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26<br />

3.2 Wavelets from filter banks . . . . . . . . . . . . . . . . . . . . . . . . . . . 27<br />

3.2.1 Perfect reconstruction . . . . . . . . . . . . . . . . . . . . . . . . . 29<br />

3.2.2 Orthogonal filter banks . . . . . . . . . . . . . . . . . . . . . . . . 29<br />

3.2.3 Multi-resolution analysis . . . . . . . . . . . . . . . . . . . . . . . . 31<br />

3.2.4 Wavelet and scaling functions . . . . . . . . . . . . . . . . . . . . . 31<br />

3.2.5 Vanishing moments . . . . . . . . . . . . . . . . . . . . . . . . . . . 33<br />

3.2.6 Linear phase . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34<br />

3.2.7 Polyphase filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . 34<br />

3.3 The stationary wavelet transform . . . . . . . . . . . . . . . . . . . . . . . 36<br />

i


ii<br />

CONTENTS<br />

3.4 Multiwavelets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40<br />

4 Analog implementation of wavelets 43<br />

4.1 Dynamic translinear systems . . . . . . . . . . . . . . . . . . . . . . . . . 43<br />

4.2 Wavelet transformations as linear systems . . . . . . . . . . . . . . . . . . 45<br />

4.3 Padé approximation of wavelet functions . . . . . . . . . . . . . . . . . . . 48<br />

4.4 L2 approximation of wavelet functions . . . . . . . . . . . . . . . . . . . . 51<br />

4.4.1 Wavelets and the L2 space . . . . . . . . . . . . . . . . . . . . . . . 51<br />

4.4.2 Parameterization . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53<br />

4.4.3 Vanishing moments . . . . . . . . . . . . . . . . . . . . . . . . . . . 55<br />

4.4.4 Obtaining a good starting point . . . . . . . . . . . . . . . . . . . 57<br />

4.5 Empirical results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64<br />

5 Orthogonal wavelet design 71<br />

5.1 Measures for the quality of a given representation . . . . . . . . . . . . . . 72<br />

5.2 Wavelet parameterization and design . . . . . . . . . . . . . . . . . . . . . 74<br />

5.2.1 Lattice structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75<br />

5.2.2 Enforcing additional vanishing moments . . . . . . . . . . . . . . . 79<br />

5.2.3 Design and optimization . . . . . . . . . . . . . . . . . . . . . . . . 82<br />

5.2.4 Experimentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87<br />

5.3 Multiwavelet parameterization and design . . . . . . . . . . . . . . . . . . 89<br />

5.3.1 Parameterization of lossless systems . . . . . . . . . . . . . . . . . 90<br />

5.3.2 Parameterization of scalar wavelets . . . . . . . . . . . . . . . . . . 95<br />

5.3.3 Parameterization of multiwavelets . . . . . . . . . . . . . . . . . . 98<br />

5.3.4 Balanced vanishing moments . . . . . . . . . . . . . . . . . . . . . 101<br />

5.3.5 Multiwavelet design . . . . . . . . . . . . . . . . . . . . . . . . . . 106<br />

6 Biomedical applications of wavelet design 109<br />

6.1 Applications of wavelet design in cardiology . . . . . . . . . . . . . . . . . 109<br />

6.1.1 Detecting the QRS complex using orthogonal wavelet design . . . 109<br />

6.1.2 QT time measurement using designed multiwavelets . . . . . . . . 113<br />

6.2 Bias field removal from MR images using wavelet design . . . . . . . . . . 118<br />

6.2.1 Radio Frequency inhomogeneities in magnetic resonance images . . 118<br />

6.2.2 Bias field removal in magnetic resonance images . . . . . . . . . . 118<br />

6.2.3 Wavelet design for RF inhomogeneity detection . . . . . . . . . . . 119<br />

6.2.4 Filtering MR images with designed wavelets . . . . . . . . . . . . . 121<br />

7 Conclusions and directions for further research 125<br />

Bibliography 129<br />

Summary 139<br />

Samenvatting 141


CONTENTS<br />

iii<br />

Curriculum Vitae 143<br />

Lists of Symbols and Abbreviations 145<br />

Index 149


Acknowledgements<br />

The BioSens project is a cooperation with the Maastricht University and the Delft<br />

University of Technology. It is financially supported by the Technology Foundation<br />

STW (project number DTC 6418), applied science division of NWO and the technology<br />

programme of the Dutch Ministry of Economic Affairs. In addition it is supported<br />

by the following research partners: Medtronic Bakken Research Center Maastricht,<br />

Medtronic Subcutaneous Diagnostics & Monitoring in Arnhem, Maastricht Instruments<br />

B.V., Twente Medical Systems International B.V. in Enschede, SystematIC design B.V.<br />

in Delft and Weijand R&D Consultancy, B.V. in Hellevoetsluis.<br />

A Ph.D. thesis is a long-term project and it is impossible to complete it without the<br />

support of many people around you. I would like to express my gratitude to all of them.<br />

My gratitude goes to my promotor and friend Ralf Peeters who was an incredible<br />

source of new ideas and made the link between the design of orthogonal wavelets and<br />

lossless systems in which he is an expert. My colleague Jordi Heijman was a great help<br />

regarding the background on cardiology.<br />

Furthermore I would like to express my gratitude to Ronald Westra, my direct colleague<br />

and the project leader of the BioSens project in Maastricht, who, together with<br />

Wouter Serdijn and Richard Houben made the BioSens project possible. As a copromotor<br />

Ronald enthusiastically proofread several sections of my thesis and helped to make<br />

them more enjoyable to read.<br />

A special word of thanks goes to my colleagues from the Delft University of Technology:<br />

Sandro Haddad and Wouter Serdijn. Our kind cooperation lead to a considerable<br />

number of publications. The trips to Delft were always enjoyable.<br />

To the industry partners. In particular Richard Houben, Jan Peuscher, Frans Smeets<br />

and Kiwi Smit who’s active participation in the users committee contributed to the<br />

success of this project.<br />

To my colleagues of the Department of Knowledge Engineering Who supported me<br />

throughout my Ph.D. time and ensured that I had enough time beside the educational<br />

load to complete my thesis.<br />

My gratitude also goes to Orazio Gambino who provided the data and supported me<br />

with the application on bias field correction.<br />

1


2 CONTENTS<br />

And last but not least to my friends, family, parents and wife. They endured a long<br />

period in which I had little time for them, but also ensured that I was able to take my<br />

distance from my thesis when necessary. Due to unforseen circumstances, handing in the<br />

full draft of my Ph.D. thesis to my promotor and sending the manuscript to the reading<br />

committee almost coincided with the date of our marriage and the birth of our daughters<br />

Romy and Sofia, respectively. However my wife Erica courageously withstood the extra<br />

stress involved.<br />

Joël Karel


Chapter 1<br />

Introduction<br />

Paraphrasing Jane Austen [9], one can state that: It is a truth universally acknowledged,<br />

that people die. Moreover, it is an established fact that among the causes of death, the<br />

cardiovascular related diseases worldwide rank as number one [123]. Technical advances,<br />

such as pacemakers, help to reduce the share of these diseases in the total number of<br />

deaths. Paramount in the future advancement of pacemakers and similar implantable<br />

devices, is the progress that can be achieved in their ability to sift, process and interpret<br />

sensor data. In this light, this work addresses two relevant aspects of new generation<br />

implantable devices: low-power implementation combined with high computing power,<br />

and efficient signal processing, using wavelets and various aspects of systems theory.<br />

1.1 Research setting<br />

For readers unfamiliar with cardiology a brief introduction is provided in Sections 2.2–<br />

2.4. One possibility of interest for decreasing the mortality among patients at risk of<br />

a cardiovascular disorder, is the implantation of a therapeutic device or a monitoring<br />

device. A classical example of such a therapeutic device is a pacemaker [122, 51], which<br />

is further discussed in Section 2.1. The fact that permanent artificial pacemakers are implanted<br />

in the human body makes it currently still impossible to recharge them, limiting<br />

the lifespan of the device to the battery lifetime.<br />

Nowadays a sensing circuit is present that controls the frequency at which a therapy is<br />

applied, contributing to the patients’ comfort. The signal processing techniques employed<br />

in production pacemakers are relatively simple [51] in general. More advanced algorithms<br />

have the clear potential to give a considerable increase in patient comfort by tuning the<br />

mode of operation to the electrophysiological state of the heart. However, since the<br />

required sensing circuit is always active [49], this introduces a new source of power<br />

consumption. The power consumption tends to increase with the complexity of the sense<br />

amplifier, and power saving techniques are therefore a relevant topic.<br />

3


4 CHAPTER 1. INTRODUCTION<br />

In the last few decades a novel signal processing technique called wavelet transforms<br />

has emerged. As is well known, the classical Fourier transform (see Section 2.6) represents<br />

a signal in terms of its frequency components. Wavelets, as discussed in Chapter 3, on<br />

the other hand represent the signal in terms of both pseudo-frequency (scale) and place<br />

(time). In addition, wavelet analysis is capable of flexibly adapting its focus, such that<br />

it narrows the time window and widens the frequency window for high frequencies, and<br />

widens the time window and narrows the frequency window for low frequencies. This<br />

is referred to as the “zoom-in” property of wavelets. Wavelets are nowadays widely<br />

used in the field of cardiac signal processing [61, 78, 83, 102, 98, 112, 87] and signal<br />

processing in general. However, for a realistic integration in implantable devices, a powerefficient<br />

implementation is essential. As argued in Chapter 4, it is opportunistic, from<br />

a power consumption perspective, to perform as many computations as possible in the<br />

analog domain. This is due to the fact that analog-to-digital converters account for a<br />

considerable proportion of the power consumption.<br />

1.2 Research questions and thesis outline<br />

Three main research questions are addressed in this thesis:<br />

1. How to approximate wavelets for the implementation in continuous-time analog<br />

systems.<br />

In earlier work [53, 50, 49, 54, 48] analog dynamic translinear systems (see Section<br />

4.1) were used as a platform to implement continuous wavelet transforms (see<br />

Section 3.1). This involved the approximation of the wavelet transform with the impulse<br />

response of a linear system as discussed in Section 4.2. A brief background on<br />

linear systems is provided in Section 2.8. Using the technique of Padé approximation<br />

the authors of [53] obtained a rational approximation of the Laplace transform<br />

of a given wavelet function. In Section 4.3 this approach, that is only suited for the<br />

approximation of a limited number of wavelet functions, is discussed in more detail.<br />

A new approach, based on L 2 approximation, that can be used for a wider range of<br />

wavelet functions is discussed in Sections 4.4 and 4.5. This approach can even be<br />

employed for the approximation of the wavelet function of discrete wavelets such<br />

as the Daubechies 3 wavelet as demonstrated in Section 4.5. The approximation of<br />

discrete wavelets is possible due to the fact that the dilation and wavelet equation,<br />

that are discussed in Section 3.2, provide a connection between discrete-time and<br />

continuous-time. This is illustrated in Figure 1.1<br />

2. How to design (multi)wavelets for an application at hand.<br />

Another important issue of interest is the choice of a wavelet basis [103]. When employing<br />

the Fourier transform the basis functions are fixed, i.e., sines and cosines.<br />

However for wavelets this is not the case and the basis needs to satisfy the conditions<br />

associated with the wavelet framework. One possibility for wavelet selection<br />

is to use a parameterized wavelet that can be tuned to the application at hand


1.2. RESEARCH QUESTIONS AND THESIS OUTLINE 5<br />

Continuous<br />

wavelets<br />

Discrete<br />

wavelets<br />

Wavelet function<br />

Dilation and wavelet eqn<br />

Orthogonal Filter<br />

bank<br />

L2 approximation<br />

Linear system<br />

approximation of<br />

wavelet<br />

transform<br />

Circuit design<br />

Design<br />

Low-power,<br />

analog DTL<br />

implementation<br />

Wavelet<br />

parameterization<br />

Prototype signal<br />

(application dependent)<br />

Applications<br />

Design criterion<br />

Sparsity<br />

Weighting<br />

Figure 1.1: Overview of the design and analog implementation of wavelets. Continuoustime<br />

wavelets can be approximated by matching the associated wavelet function with the<br />

impulse response of a linear system. This can then be implemented using the dynamic<br />

translinear circuit approach. Orthogonal discrete wavelets have an associated filter bank.<br />

Under conditions, using the dilation and wavelet equations, an associated wavelet function<br />

can be found that can be approximated. For orthogonal wavelets from filter banks a<br />

parameterization and a design criterion are available that allow for the design of wavelets.


6 CHAPTER 1. INTRODUCTION<br />

[101]. Another possibility is to design a custom wavelet as is the case in this study<br />

[69]. Earlier work on the design of wavelets matched the amplitude spectrum and<br />

the phase spectrum of a wavelet to a reference signal in the Fourier domain separately<br />

[24]. The authors of [46] describe an algorithm for the design of biorthogonal<br />

and semi-orthogonal wavelets. The design criterion comes down to the maximization<br />

of the energy in the approximation coefficients, however no clear motivation<br />

is provided by the authors. In [92] a parameterization of orthogonal wavelets, involving<br />

polyphase filters and the lattice structure as for example in [107] (see also<br />

Section 3.2), was discussed, however no clear design criterion was provided. In Section<br />

5.1 two design criteria are introduced for discrete wavelets, that can be used<br />

to measure the quality of a discrete wavelet for representing a given signal. In this<br />

context, a wavelet of “good quality” ensures that the signal’s energy is clustered<br />

in the wavelet (time-frequency) domain, i.e., it maximizes sparsity. In Section 5.2<br />

a parameterization of discrete-time orthogonal wavelets involving polyphase filterbanks<br />

and the lattice structure as in [92, 107] is discussed. The enforcement of<br />

additional vanishing moments and the use of the design criterion is discussed as<br />

well. An integral approach for orthogonal wavelet filter design is developed and<br />

presented in that section. As discussed in [68] and illustrated in Figure 1.1 it is<br />

possible to compute a wavelet function associated with the designed wavelet filter<br />

bank. If this wavelet function is sufficiently smooth, it is possible to approximate<br />

the designed wavelet in analog circuits.<br />

When multiple well-defined features in a signal need to be distinguished simultaneously,<br />

multiwavelets [76, 77, 107] can be employed. In the multiwavelet multiple<br />

waveforms are used to jointly generate a basis for the L 2 space. In Section 3.4<br />

it is discussed how these multiwavelets can be parameterized in polyphase terms<br />

and how the condition of orthogonality comes down to requiring that a system in<br />

polyphase is “lossless” [117]. In order to design multiwavelets, a parameterization of<br />

lossless systems is required, as is discussed in Section 5.3. For this parameterization<br />

in Section 5.3.1 the “tangential Schur algorithm” [55] is used. This parameterization<br />

is first provided for scalar wavelets in Section 5.3.2 and for multiwavelets in<br />

Section 5.3.3. One of the issues involved is the enforcement of a first vanishing<br />

moment which is a requirement for a valid wavelet multiresolution structure.<br />

3. To investigate the practical potential of this approach for cardiac applications.<br />

In Section 6.1 two practical applications of (multi)wavelet design with respect to<br />

cardiac signal processing are discussed. In Section 6.1.1 it is demonstrated how<br />

scalar wavelet design is useful for the detection of QRS complexes in ECG signals.<br />

In Section 6.1.2 the length of the QT interval is estimated using designed<br />

multiwavelets. In Section 6.2 the use of wavelet design for an application in 2D<br />

image analysis is discussed. This describes the removal of an artifact in magnetic<br />

resonance imaging, called the “bias field”, using specially designed wavelets [64].


Chapter 2<br />

Cardiac signal processing<br />

In the first part of this chapter a brief introduction is provided for readers unfamiliar<br />

with cardiology in Sections 2.2–2.4. One possibility of interest to decrease the mortality<br />

of patients at risk of a cardiovascular disorder, is the implantation of a therapeutic device<br />

or a monitoring device. A classical example of such a therapeutic device is a pacemaker<br />

[122, 51], as is further discussed in Section 2.1. For this research various datasets with<br />

ECG data were used as discussed in Section 2.3. In Section 2.4 it is discussed that due<br />

to pathologies various morphologies and rhythms may be exhibited in ECG signals.<br />

In the second part a short introduction to signal processing is provided in Section 2.5.<br />

Three classical transforms of interest in this work are the Fourier transform (Section 2.6),<br />

the Laplace transform (Section 2.7) and the z-transform (Section 2.7). A brief background<br />

in linear systems is given in Section 2.8, followed by an introduction to filters in<br />

Section 2.9.<br />

2.1 Introduction<br />

Cardiovascular related diseases are still the major cause of death in The Netherlands,<br />

as well as worldwide, according to WHO data [123], accounting for 33% of the total<br />

mortalities in 2004 [60]. On average this means over 45,000 death per year or over a<br />

hundred deaths per day in The Netherlands alone. However the proportion of deaths due<br />

to cardiovascular disorders is decreasing. Prevention, better health care and technological<br />

advances account for this development.<br />

Among these technological advances are new drugs, therapies, monitoring devices and<br />

implantable therapeutic and diagnostic devices. These devices measure signals that may<br />

give an indication of the state of operation of the heart. However, since these signals<br />

are often corrupted with noise, a denoising step may be required. Denoising is one of<br />

the aspects addressed by the field of Signal processing and, in general, monitoring and<br />

diagnostic devices rely heavily on signal processing techniques. For diagnostic purposes<br />

7


8 CHAPTER 2. CARDIAC SIGNAL PROCESSING<br />

it may be necessary to store the signals that have been measured. Some devices, such<br />

as implantable diagnostic and monitoring devices, may detect events in order to decide<br />

whether or not to store the signals that are being measured.<br />

The pacemaker is the most prominent implantable therapeutic device. Annually over<br />

600,000 pacemakers are implanted worldwide [122]. Early artificial pacemakers did not<br />

possess a sensing circuit and only consisted of a power source, a pulse generator and an<br />

electrode [51]. They simply delivered a fixed-rate pulse to the heart, regardless of the state<br />

of the heart. Since this type of pacemaker can compete with the spontaneous activity of<br />

the heart, it could cause arrhythmias and other undesired effects. In addition the therapy<br />

was also applied during normal operation of the heart, i.e., when it was unnecessary. A<br />

consequence of this, is that the built-in battery was depleted relatively fast. To overcome<br />

these shortcomings, demand pacemakers were developed that do take intrinsic heart<br />

activity into account (see e.g. [51]). As a result of this development, nowadays a sensing<br />

circuit is present in pacemakers. In this sensing circuit signal processing is performed as<br />

part of the decision stage to determine whether or not stimuli are needed. Therapeutic<br />

devices may be critical in order to sustain the life of the patient, therefore their decision<br />

algorithms have to be very robust. Failure of these devices may result in lawsuits and<br />

costly claims. As a result medical companies are conservative if it comes to employing<br />

new technologies [51] in implantable devices.<br />

Because a persons life depends on the correct interpretation of the signals by the<br />

pacemaker and other medical devices, it can be concluded that signal processing is of<br />

great importance to the field of cardiology.<br />

2.2 The human heart<br />

The human heart is a special organ. It has a size that is slightly bigger than a fist and it<br />

mainly consists of muscle tissue, the myocardium, that contracts up to several billions of<br />

times during a persons lifespan. The muscle tissue of which the heart consists, is unique<br />

in the human body. It is a special kind of involuntary muscle. The heart basically acts<br />

as a pump that consists of four chambers. Oxygen-poor blood enters the heart in the<br />

right atrium (RA). This blood is pumped to the right ventricle (RV). From there it is<br />

pumped through the lungs for oxygenation and this oxygenated blood reenters the heart<br />

in the left atrium (LA). From there it is pumped to the left ventricle (LV), which pumps<br />

the oxygenated blood through the rest of the body. Since the atria only pump the blood<br />

to the adjacent ventricles their tissue mass is relatively small. In addition the tissue<br />

mass of the RV is small compared to that of the LV since the RV only has to pump the<br />

blood through the lungs whereas the LV has to pump the blood through the whole body.<br />

The cells of the myocardium conduct current and if stimulated by an electrical pulse<br />

these muscle cells contract. This is called excitation-contraction coupling [14]. In their<br />

resting state these cells are said to be polarized and their electrical potential is at resting<br />

potential. If electrical activation of a cell is initiated, Na + and Ca + ions travel through<br />

ion channels in a membrane in and out the cell. This process is called depolarization.


2.2. THE HUMAN HEART 9<br />

aortic<br />

valve<br />

SA node<br />

pulmonary<br />

valve<br />

AV node<br />

RA<br />

RV<br />

LA<br />

LV<br />

atrioventricular<br />

valves<br />

Interventricular septum<br />

+ His bundle<br />

Figure 2.1: Schematic representation of the heart<br />

In the case of ventricular depolarization the Na + ions accounts for 99% of the effect.<br />

During the 150-300 ms after the heart cell is depolarized, the dynamic balance of the<br />

electrical currents maintains what is known as an action potential. This action potential<br />

ends with the repolarization of the cell, which involves moving ions into and out of the<br />

cell in the opposite direction, allowing the cell to conduct electrical impulses again. For<br />

a more detailed background on electrophysiology the reader is referred to [124].<br />

In order for the heart to effectively pump the blood through the body, the contractions<br />

of the muscle cells have to be coordinated. Special pacemaker cells control the heart rate.<br />

This process goes through a number of steps during normal activity:<br />

i) The sino-atrial (SA) node, located at the top of the RA generates an electrical<br />

impulse that propagates relatively slowly through the muscles of the atria, such<br />

that these are depolarized. The conduction of this pulse from the AV node to the<br />

LA tracts mainly over Bachmann’s bundle.<br />

ii) When this electrical pulse reaches the atrio-ventricular (AV) node after traversing<br />

from the SA node over the anterior, middle, and posterior internodal tracts, the


10 CHAPTER 2. CARDIAC SIGNAL PROCESSING<br />

propagation is delayed. During this delay the transfer of blood from the atria to<br />

the ventricles is completed.<br />

iii) The pulse travels through the His-bundle, the bundle branch and the Purkinje<br />

fibers to the endocardium of the ventricles.<br />

iv) The ventricles contract from bottom to top stimulated by a fast moving pulse.<br />

The anisotropy of the tissue is effective here; the tissue of the heart has a certain<br />

orientation and the electrical pulses travel faster along this direction.<br />

If the regular coordination in the cardiac activity is disturbed this is called an arrythmia.<br />

The excitation and conduction system consists of the SA node, the internodal tracks,<br />

Bachmann’s bundle, the AV node, the bundle of His, bundle branches and Purkinje fibers.<br />

The AV node is innervated by the autonomic nervous system, however the intrinsic rate<br />

of the SA node during normal sinus rhythm is the highest and as a result under nonpathological<br />

circumstances the SA node controls the heart rate. Certain chemicals, e.g.<br />

epinephrine, may affect the heart rate.<br />

2.3 Electrocardiogram data<br />

A number of different measurements can be conducted on the cardiovascular system.<br />

The electrical activity of the heart can be measured with the electrocardiogram (ECG or<br />

EKG). The heart sounds or vibrations can be measured as the phonocardiogram, and<br />

the pressure waves propagating through the arteries as the arterial pulse. The ECG will<br />

be discussed in this section, where the reader is referred to [100] for a more thorough<br />

discussion on this subject.<br />

2.3.1 The Electrocardiogram<br />

A signal of particular interest in this research is the ECG. The ECG is the measurement<br />

of the electrical potential between various points on or in the body. The measurement<br />

electrodes are called leads. There are various possible set-ups for lead placement. Typically<br />

a 12-lead ECG is acquired by placing the leads on the body (surface ECG). Since<br />

the heart is a 3-dimensional (3D) object it is important that the ECG provides information<br />

about the hearts activity in three (approximately) orthogonal directions [36]. The<br />

standard lead placement for the 12-lead ECG ensures that this spatial information is<br />

provided. For a more thorough discussion on lead placement the reader is referred to for<br />

example [100, 35]. An example of the 12-lead ECG plus three extra leads for the PTB<br />

database (see Section 2.3.2) is displayed in Figure 2.2.<br />

It is also possible to measure the ECG by placing an electrode directly on the tissue<br />

of the heart. One then speaks of the intracardiac ECG (IECG) [51]. The measurement<br />

of an IECG is obviously more invasive than the measurement of the surface ECG. Yet<br />

another possibility for placing leads is to implant a monitoring device under the skin that<br />

records the subcutaneous ECG.


i<br />

2.3. ELECTROCARDIOGRAM DATA 11<br />

iii<br />

avl<br />

v1<br />

v3<br />

v5<br />

vx<br />

vz<br />

0.2<br />

−0.2 0<br />

−0.4<br />

−0.6<br />

0.2<br />

−0.2 0<br />

−0.4<br />

−0.6<br />

0.4<br />

0.2<br />

−0.2 0<br />

−0.4<br />

1<br />

0.5<br />

0<br />

1.5<br />

0.5 1<br />

−0.5 0<br />

0.2<br />

−0.2 0<br />

−0.4<br />

0.2<br />

0<br />

−0.2<br />

−0.4<br />

0.4<br />

0.2<br />

−0.2 0<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

ii<br />

avr<br />

avf<br />

v2<br />

v4<br />

v6<br />

vy<br />

0<br />

−0.2<br />

−0.4<br />

−0.6<br />

0.4<br />

0.2<br />

0<br />

0<br />

−0.2<br />

−0.4<br />

−0.6<br />

1<br />

0.5<br />

0<br />

1<br />

0.5<br />

0<br />

−0.5<br />

0.2<br />

0<br />

−0.2<br />

0.2<br />

0<br />

−0.2<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

1000 2000 3000 4000<br />

Figure 2.2: Excerpt of the ECG of patient 1 from the PTB database over 4500 ms. The<br />

conventional 12 lead ECG is displayed (leads i, ii, iii, avr, avl, avf, v1, v2, v3, v4, v5, v6)<br />

along the three Frank leads (vx, vy, vz). The ECG was digitized at 1 kHz. In this ECG,<br />

right-bundle branch block is visible.


12 CHAPTER 2. CARDIAC SIGNAL PROCESSING<br />

QT<br />

ST<br />

P<br />

PQ<br />

QRS<br />

T<br />

U<br />

Figure 2.3: Segments of the ECG<br />

From the ECG recordings various waves can be distinguished that correspond to<br />

the propagation of electrical activity across the heart as described in Section 2.2. The<br />

morphology and variability of these waves can vary from lead to lead due to the fact that<br />

the heart is a 3D organ. Also the muscle mass is not uniform throughout the heart, and as<br />

a consequence it is easier to see the activity of the ventricles than of the atria. The leads<br />

may also pick up signals from other sources, such as respiration, muscle movement and<br />

electrode movement. Respiration can cause low-frequency noise. But muscle movement,<br />

and to a lesser extent electrode movement, may cause in-band noise which can be much<br />

harder to filter out. The heart may also suffer from an arrythmia that gives rise to a<br />

different morphology. A number of examples of morphologically different ECG waves is<br />

displayed in Figure 2.4.<br />

Due to the step-wise operation of the heart during normal sinus rhythm, a number<br />

of physiological meaningful segments can be identified in the ECG (see for example [59])<br />

as displayed in Figure 2.3:<br />

P wave Depolarization of the atria. Corresponds to step i on page 10.<br />

PQ segment Propagation delay in the AV node (step ii) on page 10) corresponds to<br />

iso-electric segment.<br />

QRS complex Ventricular depolarization (steps iii and iv on page 10). Due to the<br />

muscle mass of the ventricles this generally is the dominant complex. Meanwhile<br />

atrial repolarization occurs.<br />

ST segment Ventricular muscle cells maintain their action potential resulting in an<br />

iso-electric segment.<br />

T wave Ventricular repolarization.<br />

Sometimes other small amplitude waves such as the “U wave” are visible [59].<br />

A number of intervals of interest can be derived from these segments:


2.3. ELECTROCARDIOGRAM DATA 13<br />

1200<br />

1100<br />

1000<br />

1200<br />

1100<br />

1000<br />

1100<br />

1050<br />

1000<br />

950<br />

50 100 150 200 250<br />

50 100 150 200 250<br />

50 100 150 200 250<br />

1300<br />

1200<br />

1100<br />

1000<br />

1200<br />

1100<br />

1000<br />

1100<br />

1000<br />

900<br />

800<br />

700<br />

50 100 150 200 250<br />

50 100 150 200 250<br />

50 100 150 200 250<br />

1200<br />

1100<br />

1000<br />

1100<br />

1000<br />

900<br />

1200<br />

1100<br />

1000<br />

900<br />

50 100 150 200 250<br />

50 100 150 200 250<br />

50 100 150 200 250<br />

Figure 2.4: Examples of smoothed ECG beats from the MIT-BIH arrythmia database.<br />

QT interval Duration of ventricular depolarization and repolarization. Also indicated<br />

as the duration of the ventricular electrical systole, which is the electrical activity<br />

that stimulates the ventricles to contract. During the diastole the myocardium<br />

relaxes.<br />

RR interval Duration of a complete ventricular cardiac cycle.<br />

PP interval Duration of a complete atrial cardiac cycle.<br />

Due to the physical meaning of these time segments their proper identification and<br />

duration is essential for the assessment of the patients condition. In addition the various<br />

segments may have a certain morphology. The duration of the segments and their<br />

morphology may be related to a certain rhythm; that may in turn correspond to a certain<br />

arrythmia. In clinical situations a cardiologist judges the ECG, and using his/her<br />

expert knowledge, determines the rhythm. However this decision is a challenging task<br />

for automated systems. In particular if the requirement is that the system has to work<br />

in real-time, and especially if it concerns an implantable device.


14 CHAPTER 2. CARDIAC SIGNAL PROCESSING<br />

2.3.2 Signal archives<br />

The main source of data used in this study are the Physionet signal archives [43]. This<br />

is a collection of databases of various types of signals, of which ECG recordings form the<br />

majority. The most prominent recordings are the MIT-BIH databases that were recorded<br />

by the Beth Israel Deaconess Medical Center and Massachusetts Institute of Technology.<br />

In particular the MIT-BIH arrythmia database is widely used for both the evaluation of<br />

arrhythmia detectors as well as for basic research into cardiac dynamics. Additionally<br />

the Physikalisch-Technische Bundesanstalt (PTB) database [16] which is also available<br />

through Physionet has been used in this research. This same database was used for<br />

the Computers in Cardiology 2006 Challenge [88] on QT interval estimation. One of<br />

the participants of this challenge developed a set of manual annotations that has been<br />

published in [25].<br />

In addition datasets from industry partners were available. Medtronic supplied a<br />

number of datasets that were recorded with a Medtronic RevealR subcutaneous leadless<br />

ECG recorder.<br />

2.4 Cardiac rhythms and pathologies<br />

From the morphology and frequency of the various complexes in the ECG, cardiologists<br />

can determine the involved cardiac rhythm. Furthermore the patient may suffer<br />

from a certain pathology that is associated with the rhythm at hand, or with a specific<br />

morphology of one of the complexes in the ECG. In order to arrive at more effective therapies<br />

the diagnostic capabilities of therapeutic devices have to be improved. Detecting<br />

morphologies in ECGs may be of great importance for expanding these capabilities.<br />

The human heart is a syncytium, which means that electrical pulses can propagate<br />

freely from cell to cell in any direction on the myocardium. In normal sinus rhythm this<br />

propagation is done in an orderly fashion. However if this process is disturbed, turbulent<br />

and disorganized propagation of the waves causes the myocardium to contract in a chaotic<br />

manner; it fibrillates. This can occur in the atria, hence atrial fibrillation (AF). Often<br />

AF is asymptomatic, i.e., the patient does not experience any grief from this arrythmia.<br />

However, there is an elevated risk of stroke since due to the poor circulation of blood<br />

during AF the blood may pool and clot. Ventricular fibrillation (VF), puts the patient<br />

in acute risk of sudden cardiac death. In order to restore normal sinus rhythm in case<br />

of VF, a defibrillator is used which generates a large electrical discharge to “reset” the<br />

heart. For patients at chronic risk of VF, a cardioverter-defibrillator can be implanted<br />

that monitors the heart, and in case an arrhythmia is detected, therapy is applied.<br />

There are a large number of arrhythmias and pathologies that can occur [13]. For<br />

effective diagnosis and treatment, signal processing is a very important tool, certainly in<br />

the absence of a physician when automated diagnosis is required such as in implantable<br />

medical devices.


2.5. SIGNAL PROCESSING 15<br />

2.5 Signal processing<br />

Signal processing is the field that aims to analyze, manipulate and interpret signals.<br />

This includes the removal of noise, separation of signals from various sources, feature<br />

extraction, storage (optionally with compression) and signal representation.<br />

Signals can be either continuous-time or discrete-time, with either analog or digital<br />

values. It is popular convention not to make the distinction between whether the<br />

signal is sampled (continuous/discrete-time) and how these samples are quantified (analog/digital).<br />

Instead “analog” is a popular term for continuous-time analog signals and<br />

“digital” is a popular term for discrete-time digital signals. In this work the terms “analog”<br />

and “digital” will refer to the representation of the signal values, regardless whether<br />

they are sampled or not.<br />

Most sensor information concerns continuous-time analog signals with the signal values<br />

represented in voltage or in ampères. In order to obtain a digital signal, the sensor<br />

data has to be quantified. This is done by an analog-to-digital (A/D) converter. This<br />

converter has a certain number of bits that together with the range of both the converter<br />

and the signal determines the precision it can obtain.<br />

The A/D converter usually also samples the input signal such that the output signal<br />

becomes a discrete-time digital signal. The number of samples per time unit is the sampling<br />

rate or the sampling frequency. The Nyquist frequency is a property of a discretetime<br />

system and defined as half the sampling frequency. The Nyquist rate [94, 104] is twice<br />

the highest frequency that can be perfectly reconstructed in a continuous-time signal,<br />

using interpolation, from a discrete-time signal which has been sampled at a frequency<br />

equal to this Nyquist rate. So if the Nyquist frequency (half of the sampling frequency)<br />

of the sampled system is larger than the bandwidth (half of the Nyquist rate) of the<br />

continuous-time source signal, then perfect reconstruction should, at least in theory, be<br />

possible.<br />

Signals can be represented in various domains. Sensor information such as ECGs is<br />

usually represented in the time domain. If one wants to examine certain characteristics of<br />

the signal it can be opportunistic to examine the signal in a different domain. For example<br />

in order to view the frequency content of a signal one can use the frequency domain, or<br />

if one wants to analyze time information and frequency information simultaneously, one<br />

can use the wavelet domain. To convert a signal to the frequency domain or to the<br />

wavelet domain, the Fourier (or Laplace) transform or the wavelet transform can be<br />

used, respectively.<br />

2.6 Fourier transforms<br />

The Fourier transform is a linear operator that maps a finite energy time domain function<br />

f(t) to a frequency domain function F (ω) [15, 17]. The function f(t) is expressed in terms<br />

of F (ω) as:<br />

f(t) = √ 1 ∫ ∞<br />

F (ω)e iωt dω, (2.1)<br />

2π<br />

−∞


16 CHAPTER 2. CARDIAC SIGNAL PROCESSING<br />

where ω is the angular frequency variable (in radians per second) and i is the complex<br />

number. This formula is used in the “synthesis”.<br />

The Fourier transform (used in the “analysis”) is defined as:<br />

F (ω) = 1 √<br />

2π<br />

∫ ∞<br />

−∞<br />

e −iωt f(t)dt, (2.2)<br />

Note that the product of the normalization factors of (2.2) and (2.1) has to be equal to<br />

1<br />

2π<br />

. In this case the normalization factors are chosen to be symmetrical.<br />

A time domain signal f(t) is said to have finite energy if its L 2 -norm is finite:<br />

√ ∫ ∞<br />

||f|| 2 = |f(t)| 2 dt < ∞ (2.3)<br />

−∞<br />

The angular frequency ω is related to the frequency h in Hertz (Hz) by the identities<br />

ω = 2πh and h = ω 2π . (2.4)<br />

When F (ω) is expressed in polar coordinates as F (ω) = r(ω)e iφ(ω) , further meaning<br />

can be given to the complex function F (ω) (2.1). The function r(ω) represents the<br />

magnitude, i.e. the amplitude, of the complex harmonics e iωt per frequency, and φ(ω)<br />

their initial phase angles. This can be expressed in a pair of Bode plots. The Bode<br />

magnitude plot shows the logarithm of the magnitude in relation to the frequency, plotted<br />

on a logarithmic frequency axis. The frequency response of a linear, time-invariant system<br />

(see Section 2.8) can for example be visualized in such manner. The Bode phase plot<br />

shows the initial phase angle in relation to the frequency, again plotted on a logarithmic<br />

frequency axis.<br />

Using Euler’s formula (see e.g. [39])<br />

e iφ = cos φ + i sin φ, (2.5)<br />

it becomes clear that the Fourier transform decomposes the signal in terms of (orthogonal)<br />

sine and cosine basis functions.<br />

If the Fourier transformed signal has compact support in the sense that there exists<br />

an angular frequency ω 0 such that:<br />

F (ω) = 0, ∀|ω| > ω 0 , (2.6)<br />

then the signal is said to be bandlimited with bandwidth ω 0 .<br />

If the function f is periodic then it will obviously have infinite energy. However the<br />

power, i.e., the energy per time unit, can still be required to be finite. The Fourier<br />

transform over a single period will then be calculated and one then sometimes speaks of<br />

a Fourier series. However it is popular convention to reserve this term for the discrete<br />

Fourier transform of a periodic function.


2.6. FOURIER TRANSFORMS 17<br />

Of the many interesting properties of the Fourier transform, one of particular interest<br />

is that convolution in the time domain of two functions f and g<br />

h(t) = (f ∗ g)(t) =<br />

∫ ∞<br />

−∞<br />

becomes multiplication in the Fourier domain:<br />

f(τ)g(t − τ)dτ, (2.7)<br />

H(ω) = √ 2πF (ω)G(ω), (2.8)<br />

where F (ω), G(ω) and H(ω) are the Fourier transforms of f(t), g(t) and h(t) respectively.<br />

This property is particularly useful in the field of signal processing where convolution<br />

is extensively used. The Fourier transform has many more properties. For more details<br />

the reader is referred to for example [15, 84].<br />

When determining the discrete Fourier transform (analysis) X of a (possibly complex)<br />

discrete-time signal x = {x[0], . . . , x[N − 1]}, one treats this finite length signal x as<br />

if it were periodic. If the complex harmonics e −iΩn that are used as a basis for the<br />

decomposition are chosen such that an integer number of periods fit in the length of the<br />

signal x, this indeed gives the desired result. One then speaks of the discrete Fourier<br />

series. The discrete Fourier series is defined as:<br />

X[k] = √ 1<br />

N−1<br />

∑<br />

x[n]e −iΩn , (2.9)<br />

N<br />

where Ω = 2πk/N, k = 0, 0, . . . , N − 1. Its inverse transform (synthesis) is given by:<br />

n=0<br />

x[n] = √ 1<br />

N−1<br />

∑<br />

X[k]e iΩn . (2.10)<br />

N<br />

k=0<br />

For a pair of vectors (e.g. finite length discrete-time signals) a and b the convolution<br />

product c = a ∗ b is defined as the vector:<br />

c[n] = (a ∗ b)[n] =<br />

n∑<br />

a[l]b[n − l], n ∈ {0, 1, . . . , N − 1}. (2.11)<br />

l=0<br />

In the Fourier domain this again becomes multiplication:<br />

C[k] = √ NA[k]B[k], (2.12)<br />

where A, B and C are the Fourier series of a, b and c respectively.<br />

Since the basis functions of the Fourier transform are sinusoids that extend over<br />

the entire real line R, the Fourier transform is not well equipped for studying transient<br />

phenomena. A classical way around this issue is to use a windowing function w(t) to<br />

arrive at:<br />

F w (τ, ω) = √ 1 ∫ ∞<br />

w(t − τ)e −iωt f(t)dt, (2.13)<br />

2π<br />

−∞


18 CHAPTER 2. CARDIAC SIGNAL PROCESSING<br />

This transform is called the windowed Fourier transform alias the short time Fourier<br />

transform [40]. The time window w(t − τ) localizes the transform around time τ and the<br />

exponential e −iωt in frequency around frequency ω. Due to the Heisenberg uncertainty<br />

principle [58] from the field of quantum dynamics, one cannot obtain an arbitrary accurate<br />

time-frequency measurement. Instead the product of the “standard deviations”<br />

of the time- and frequency windows, i.e., the area of the uncertainty rectangle, obeys a<br />

certain lower bound: σ τ σ ω ≥ 1 2<br />

. This area is for instance minimal if w(t) is a Gaussian<br />

function, i.e., in the case of the Gabor transform. For the windowed Fourier transform<br />

the uncertainty rectangle has constant dimensions and a variable placement in the timefrequency<br />

plane (see for example [80]).<br />

2.7 Laplace and z-transforms<br />

The (one-sided) Laplace transform (see e.g. [62]) is conveniently introduced for functions<br />

y(t) defined only for t ≥ 0. It is a function of a complex variable s and given by:<br />

Y (s) = L{y(t)} =<br />

∫ ∞<br />

t=0<br />

y(t)e −st dt. (2.14)<br />

Formally, it is defined for values of s for which the integral in (2.14) converges, which<br />

constitutes the region of convergence. This usually is a half-plane. An important class<br />

of functions used in signal processing and filtering involves sines, cosines, exponentials,<br />

polynomials and from the non-standard functions the impulse functions, all of which<br />

have rational Laplace transforms [62].<br />

The continuous-time impulse function is the Dirac delta-function δ(t) is given by:<br />

{<br />

δ(t) = 0, ∀t ≠ 0<br />

∫ ∞<br />

−∞ δ(t)dt = 1 (2.15)<br />

yielding an infinitesimally narrow, infinitely tall pulse which integrates to unity. On the<br />

other hand the discrete-time impulse can be associated (via zero-order hold (ZOH)) with<br />

a continuous time block function, called the Kronecker delta function:<br />

{ δ[n] = 0, ∀n ≠ 0<br />

(2.16)<br />

δ[n] = 1, for n = 0.<br />

Writing s = σ + iω, it is clear that the Fourier transform is obtained for σ = 0:<br />

Y (iω) = Y (ω). (2.17)<br />

(Laplace)<br />

(Fourier)<br />

Note that with abuse of notation we denote the Laplace transform of y(t) as Y (s), the<br />

Laplace transform for s = iω as Y (iω), and the Fourier transform of y(t) as Y (ω).<br />

As is well-known, for the Laplace transform Y (s) of a function y(t) the initial value<br />

theorem:<br />

y(0) = lim sY (s), (2.18)<br />

|s|→∞


2.8. LINEAR SYSTEMS 19<br />

and the final value theorem:<br />

lim y(t) = lim sY (s), (2.19)<br />

t→∞ s→0<br />

hold.<br />

In discrete time, a similar transform which is called the z-transform is used. Like<br />

s for the Laplace transform, z is a complex variable. The z-transform of an (infinite)<br />

sequence y[k], k ∈ Z is denoted by Z{y k } = Y (z) and is given by:<br />

Y (z) = Z{y k } = ∑ k<br />

y k z −k . (2.20)<br />

Observe the similarity with the definition of the Laplace transform in (2.14). The z-<br />

transform is the Laplace transform of an impulse train of a continuous function, with<br />

sampling period T ∈ R + . Given this instantanious sampled function:<br />

∞∑<br />

n=−∞<br />

y(nT )δ(t − nT ), (2.21)<br />

where δ(k) is the Dirac delta as in (2.7). One can take the Laplace transform of this<br />

sampled function and substitute z = e sT to obtain the z-transform of y(t). When taking<br />

z = e iΩ in (2.20) the Fourier series as in (2.9) is obtained.<br />

2.8 Linear systems<br />

Let S be a linear system with input u and output y. Then the system S has the following<br />

defining properties:<br />

Homogeneity If the input u to a linear system S is scaled with a factor α, the output<br />

of that system is scaled with a factor α as well: S(αu) = αS(u).<br />

Superposition If two inputs are added and passed through a linear system, then the<br />

output equals the sum of the outputs as if the two input signals were passed through<br />

the system individually: S(u 1 + u 2 ) = S(u 1 ) + S(u 2 ).<br />

Linear systems can be discrete-time or continuous-time, but in this section the focus<br />

will be on continuous-time systems. Linear systems can also be time-invariant: for any<br />

input u producing a corresponding output y = S(u), it holds that y τ = S(u τ ) for any<br />

time shift τ, where u τ , y τ denote the signals u, y time shifted by τ. If a system is<br />

both linear and time-invariant, one calls it a linear time-invariant system or LTI system<br />

for short. The input/output relation of LTI systems can be represented by differential<br />

equations. As an example of such an n th order differential equation takes the following<br />

form:<br />

y (n) (t) + a 1 y (n−1) (t) + . . . + a n−1 y (1) (t) + a n y(t)<br />

= b 0 u (n) (t) + b 1 u (n−1) (t) + . . . + b n−1 u (1) (t) + b n u(t), (2.22)


20 CHAPTER 2. CARDIAC SIGNAL PROCESSING<br />

where y (k) (t) denotes the k th derivative of y(t) with respect to t.<br />

A system is causal if any input u with u(t) = 0, ∀t < 0 produces an output y for<br />

which it holds that y(t) = 0, ∀t < 0.<br />

LTI systems have an associated impulse response, step response and transfer function,<br />

each characterizing the system. For a detailed discussion on these topics the reader is<br />

referred to for example [62]. The impulse response function h(t) is defined as the output<br />

of a system, corresponding to the Dirac delta δ(t) (2.15) as an input for continuous-time<br />

systems, and the Kronecker delta δ[n] (2.16) for discrete-time systems.<br />

The transfer function the Laplace transform of the impulse response. It is well known<br />

that in the case of zero initial conditions the following relation holds for the Laplace<br />

transform of the input U(s), the Laplace transform of the output Y (s) and the transfer<br />

function H(s) of the LTI system:<br />

Y (s) = H(s)U(s). (2.23)<br />

The Laplace transform of δ(t) is: L {δ(t)} = D(s) = 1 so that for the Laplace transformed<br />

output it holds that:<br />

Y (s) = H(s)D(s) = H(s). (2.24)<br />

If a causal LTI system is of finite order, it will possess a proper rational transfer function<br />

[62] H(s):<br />

H(s) = Y (s)<br />

U(s) = b 0s n + b 1 s n−1 + . . . + b n<br />

s n + a 1 s n−1 . (2.25)<br />

+ . . . + a n<br />

The roots of the numerator polynomial in (2.25) are called the zeros of the system and<br />

the roots of the denominator polynomial the poles. The order n of the system is defined<br />

as the McMillan degree of H(s). For single-input single-output (SISO) systems, the order<br />

of the system equals the degree of the denominator of the transfer function H(s), after<br />

canceling all common factors. In this work linear systems will be assumed to be finite<br />

order, time-invariant and causal, unless stated otherwise.<br />

Given a system with impulse response function h(t) and input u(t), the output y(t)<br />

becomes the convolution of the input u(t) and the impulse response h(t):<br />

y(t) = (h ∗ u)(t) =<br />

∫ ∞<br />

−∞<br />

h(τ)u(t − τ)dτ. (2.26)<br />

The impulse response function is the derivative of the step response function, and therefore<br />

this latter one can be calculated from the transfer function as:<br />

L −1 { 1<br />

s H(s) }<br />

. (2.27)<br />

An LTI system is said to be asymptotically stable if the output corresponding to an<br />

impulse input damps out to zero. Formally this comes down to the requirement that the<br />

poles of its transfer function are all in the open complex left halfplane.<br />

LTI systems can be represented in state-space:<br />

ẋ(t) = Ax(t) + Bu(t)<br />

y(t) = Cx(t) + Du(t).<br />

(2.28)


2.8. LINEAR SYSTEMS 21<br />

The vector x(t) is the state vector. The n × n matrix A is called the system matrix, the<br />

dynamical matrix or the state matrix, the n × 1 vector B is called the input vector, the<br />

1 × n vector C the output vector and the scalar D is called the direct feedthrough term,<br />

alias direct current or DC-term. A system may be denoted by this quadruple of matrices<br />

S = (A, B, C, D).<br />

We usually consider systems as input/output systems, hence focusing on its external<br />

interactions. The system is identified with its transfer function. Its state-space representations<br />

is merely used as a convenient “tool”, and having meaningful states in this<br />

representation is not a requirement. To arrive at computationally convenient canonical<br />

forms and parameterizations, similarity transforms are used. Given any non-singular<br />

matrix T , a similarity transformation from a basis x(t) to a basis ˆx(t) = T x(t) may be<br />

performed yielding an equivalent system:<br />

where<br />

˙ˆx(t) = Âx(t) + ˆBu(t)<br />

y(t) = Ĉx(t) + Du(t), (2.29)<br />

 = T AT −1 (2.30)<br />

ˆB = T B (2.31)<br />

Ĉ = CT −1 (2.32)<br />

A canonical form is a “standard” form. We deal with a canonical form if for all<br />

systems that are mutually equivalent to some given system, we transform it to the same<br />

member of that equivalence class. As a result, in a canonical form different choices of the<br />

parameters produce systems that are never equivalent. Using similarity transformations,<br />

systems can be brought into a canonical form and from one canonical form to another.<br />

For state-space systems the order n has the interpretation that it is the state-space<br />

dimension that is minimally required to represent the system.<br />

From the state-space representation S = (A, B, C, D) the transfer function can be<br />

determined:<br />

H(s) = D + C(sI − A) −1 B. (2.33)<br />

The poles of the system correspond to the eigenvalues of the system matrix A. As<br />

one would expect from the above it is also possible to compute the impulse response<br />

associated with the system:<br />

h(t) = Ce At B + Dδ(t), t ≥ 0. (2.34)<br />

Insted of a SISO system a multi-input multi-output system (MIMO) is considered<br />

with p inputs and q outputs. The vectors B and C and the scalar D change accordingly.<br />

B then becomes an n × p input matrix, C a q × n output matrix and D a q × p direct<br />

feedthrough. All other formulas previously discussed in this section still apply.


22 CHAPTER 2. CARDIAC SIGNAL PROCESSING<br />

For discrete-time systems similar relations and properties as for continuous-time systems<br />

hold. The state-space description in (2.28) takes the form:<br />

x[k + 1] = Ax[k] + Bu[k]<br />

y[k] = Cx[k] + Du[k].<br />

(2.35)<br />

The transfer function can be obtained from the state-space representation similarly as in<br />

(2.33) as<br />

H(z) = D + C(zI − A) −1 B, (2.36)<br />

and (2.34) changes to:<br />

h[k] =<br />

{ D k = 0<br />

CA k−1 B k > 0.<br />

(2.37)<br />

For a more in-depth discussion on linear systems and their properties the reader is<br />

referred to the relevant literature, for example [62, 115].<br />

2.9 Filtering signals<br />

It is well-known that a sinusoid u s (t) = sin(ωt) used as input for a stable linear system<br />

with transfer function H(s) yields a steady state output y s (t) as:<br />

y s (t) = A sin(ωt + φ), (2.38)<br />

where A = |H(iω)| is that absolute value and φ = ∠H(iω) is the angle of the complex<br />

number H(iω) in the complex plane. The function H(iω) is called the frequency response<br />

function of the system and is the Fourier transform of the impulse response h(t). Equation<br />

(2.38) is called the property of sinusoidal fidelity. This property implies that if a sinusoid<br />

is used as an input to such a system, then a sinusoid of the same frequency will constitute<br />

the output. However the output sinusoid exhibits a phase shift φ relative to the input<br />

sinusoid and an amplitude gain of |H(iω)|. If the phase response of the LTI system is<br />

linearwith the frequency, it is said that the system has linear phase. In a linear phase<br />

system all frequencies have equal delay times, so that there is no phase distortion.<br />

As discussed in Section 2.6, each signal can be decomposed in terms of sines and<br />

cosines. Consequently, if a Bode plot of the frequency response of the system is considered,<br />

one can examine how each frequency in an input u(t) signal is affected by the<br />

system. How each frequency is affected by the system depends on the placement of the<br />

poles and zeros in the complex s-plane: frequencies near zeros will be suppressed and<br />

frequencies near poles will be amplified. However, if poles and zeros are sufficiently close<br />

there will be interaction between the poles and zeros. By appropriate placement of the<br />

poles different type of filters can be created such as low-pass, high-pass, band-pass and<br />

notch filters. These can have properties such as fast roll-off, no pass/stop-band ripple,<br />

etc. Classically, these filters are designed to have certain properties in the Fourier<br />

domain.


2.9. FILTERING SIGNALS 23<br />

It is useful to consider filtering in the time-domain. From (2.26) it is clear that if<br />

a signal u(t) is used as an input, it is convoluted with the impulse response h(t) of the<br />

system, which is the inverse Laplace transform of the transfer function.<br />

A similar property holds in discrete time. For a discrete-time signal that is a sampled<br />

version of a continuous-time signal, as discussed on page 15, meaning can be given to<br />

the sampling frequency f in Hertz. Since sensor data is mostly continuous-time data, for<br />

most applications this f is important to relate to the frequencies in the source signal. For<br />

applications where the data is inherently discrete-time one can choose f = 1. Given an<br />

(<br />

input u s [k] = sin<br />

H(z) is:<br />

ωk<br />

f<br />

)<br />

, the corresponding output y s [k] of a system with transfer function<br />

[ ] ωk<br />

y s [k] = A sin<br />

f + φ , (2.39)<br />

where where A = |H ( e ) iω/f | and φ = ∠H ( e ) iω/f . The function H ( e ) iω/f is the<br />

discrete-time frequency response. Frequencies in input signals are thus similarly affected<br />

as in the continuous-time case.<br />

As in (2.20) let u(z) = u 0 + u 1 z −1 + u 2 z −2 + . . . be the z-transform of a discretetime<br />

signal u = {u 0 , u 1 , u 2 , . . .}. For discrete-time filters with a proper rational transfer<br />

function of the following form:<br />

H(z) = b 0z n + b 1 z n−1 + . . . + b n<br />

z n = b 0 + b 1 z −1 + . . . + b n z −n , (2.40)<br />

the z-transformed output corresponding to a transformed input u(z) can be computed<br />

as:<br />

H(z)u(z) = b 0 u 0 + (b 0 u 1 + b 1 u 0 )z −1 + (b 0 u 2 + b 1 u 1 + b 2 u 0 )z −2<br />

+ (b 0 u 3 + b 1 u 2 + b 2 u 1 + b 3 u 0 )z −3 . . . (2.41)<br />

The impulse response u 0 = 1, u k = 0 for k > 0 can be easily read off from this expression<br />

yielding {b 0 , b 1 , . . . , b n , 0, 0, . . .} and will have a finite length, i.e., h[k] = 0, ∀k > n. This<br />

system is said to be of finite impulse response (FIR). Since in linear systems the input<br />

is convoluted with the impulse response, such systems exhibit a convenient one-to-one<br />

correspondence of the transformation from input to output in the time and z-domain.<br />

The output at each time instance is the weighted mean of the input around that instance,<br />

where the weighing factors are defined by the numerator polynomial of the transfer<br />

function, hence the name moving average (MA) filters. The filter banks associated with<br />

the discrete wavelet transform in Section 3.2 take such a form (2.40).<br />

Historically filters are designed based on frequency domain properties. However in<br />

the case of in-band noise, such as muscle artifact noise in ECGs, the signal is not effectively<br />

separated from the noise in the frequency representation and consequently the<br />

noise cannot be removed using this approach. However the noise may have a different<br />

morphology from that of the signal and filtering techniques that take morphology into<br />

account, such as wavelets, may be beneficial.


Chapter 3<br />

Wavelet transformations<br />

Nowadays, a quarter of a century after Morlet and Grossman first coined the word<br />

“wavelet” and formalized the corresponding transform [44], wavelets are recognized as a<br />

fundamental and powerful tool in signal analysis, and widely applied in practice, including<br />

in the field of biomedical engineering. The remarkable rise and success of wavelets,<br />

starting from the 70s, may largely be attributed to its added value to Fourier analysis.<br />

The wavelet transform offers both time and frequency localization, whereas the Fourier<br />

transform only offers frequency information, making the wavelet transform a powerful<br />

tool for signal analysis [79, 107]. The windowed Fourier transform (2.13) offers the<br />

ability to obtain both time and frequency information too, but this method does not<br />

have the flexibility of the wavelet transform since the basis is restricted to sines and<br />

cosines (harmonics) and it does not possess the “zoom-in” property of wavelets as discussed<br />

on page page 26. In the late 1980s Yves Meyer published work on orthogonal<br />

wavelets (e.g.[85]). This inspired Stéphane Mallat to construct a multi-resolution theory<br />

on wavelets [79], whereas Ingrid Daubechies constructed a well-known set of orthonormal<br />

wavelets [30]. Despite its recent invention, wavelet theory is already well established and<br />

employed in many standards such as the jpeg standard for image compression.<br />

Many biomedical applications have been developed around wavelets. For example<br />

the detection of characteristic points in ECGs [105, 61, 78, 83, 102, 2], filtering [98, 112],<br />

separation of fetal cardiac activity [87] and many others. In [101] it has been shown<br />

that wavelets can be finetuned for a given application. Due to the work in [107, 92]<br />

on orthogonal filter banks a parameterization is available for scalar wavelets that can be<br />

used as a framework for wavelet design. However no optimization criterion was discussed.<br />

In Chapter 5 an optimization criterion is introduced for the design of optimal wavelets.<br />

In this chapter we will briefly introduce some basic concepts of wavelets, and then<br />

focus in some depth on the relation between wavelets and filter banks. This chapter will<br />

close with a discussion on multiwavelets and their relation with lossless systems.<br />

25


26 CHAPTER 3. WAVELET TRANSFORMATIONS<br />

3.1 Continuous wavelets<br />

The wavelet transformation if aimed at the decomposition of a signal, localized in both<br />

time and frequency simultaneously by inducing a change of basis for the signal in question.<br />

In order to accomplish this a so called “wavelet” function is used as a basis, and this<br />

wavelet function has the property that it is localized in both time and frequency. More<br />

formally, the wavelet transformation of a real signal x(t) using a real wavelet ψ(t) is<br />

defined as the cross-covariance at lag τ of that signal with the normalized dilated wavelet<br />

1 √ σ<br />

ψ( t σ<br />

) (see for example [44, 80]):<br />

W (τ, σ) =<br />

∫ ∞<br />

−∞<br />

x(t) 1 √ σ<br />

ψ<br />

( t − τ<br />

σ<br />

)<br />

dt, τ ∈ R, σ ∈ R + . (3.1)<br />

The parameter σ is a scale parameter that determines the (pseudo)frequency localization<br />

of the wavelet transform and the parameter τ determines the time localization of the<br />

wavelet transform. For a specific scale σ and a specific time τ the L 2 -inner product<br />

1<br />

between the signal x(t) and the normalized, time-shifted and scaled wavelet √σ ψ ( )<br />

t−τ<br />

σ<br />

is thus computed. The form in (3.1) is known as the continuous wavelet transform (CWT)<br />

or as the integral wavelet transform. The continuous wavelet transform is an invertible<br />

transform as shown below:<br />

x(t) = 1 ∫ ∞ ∫ ∞<br />

W (τ, σ) 1 ( ) t − τ<br />

√ ψ dτ dσ<br />

C<br />

σ σ σ 2 , t ∈ R, C ∈ R+ , (3.2)<br />

0<br />

−∞<br />

where C is a constant that depends on the wavelet ψ(t) at hand. Not any arbitrary function<br />

can be used as a wavelet basis and therefore the function ψ(t) must obey a so-called<br />

admissability condition [44]. The admissibility condition comes down to requiring that<br />

ψ(t) has zero average and that its Fourier transform Ψ(ω) is continuously differentiable<br />

[80]. For a more detailed discussion on this subject the reader is referred to for example<br />

[80, 26].<br />

We denote dilated and time-shifted wavelet functions as<br />

ψ τ,σ (t) = 1 √ σ<br />

ψ<br />

Their corresponding Fourier transform is:<br />

( t − τ<br />

σ<br />

)<br />

. (3.3)<br />

Ψ τ,σ (ω) = e −iτω Ψ(σω). (3.4)<br />

From (3.4) it can be seen that unlike in (2.13) where the Heisenberg uncertainty rectangle<br />

has constant dimensions in the time-scale plane. In this case when moving to the coarser<br />

scales (i.e. lower frequencies) the window tightens in the scale dimension and widens in<br />

the time dimension. Conversely the window widens in the scale dimension and tightens<br />

in the time dimension when moving to finer scales (higher frequencies). The dimensions<br />

of the uncertainty rectangle remain constant when moving it along the time dimension.<br />

This behavior of the uncertainty rectangle facilitates the “zoom-in” property of wavelets.


3.2. WAVELETS FROM FILTER BANKS 27<br />

0.8<br />

0.6<br />

0.4<br />

Gaussian wavelet<br />

Mexican Hat wavelet<br />

0.2<br />

0<br />

−0.2<br />

−0.4<br />

−0.6<br />

−0.8<br />

−1<br />

−5 0 5<br />

Figure 3.1: The Gaussian and the Mexican Hat wavelet<br />

In cardiac signal processing the Gaussian wavelet ψ(t) = −2( 2 π ) 1 4 te −t2 is very popular<br />

(see for example [5, 102]), partly due to the fact that the wave-like shape as shown in<br />

Figure 3.1 resembles the QRS complex, partly due to some of its other nice properties<br />

such as the fact that it is a derivative of a smoothing function [80]. The Gaussian wavelet<br />

is constructed by taking the normalized derivative of the Gaussian probability density<br />

function (pdf) with mean zero and variance 0.5. Higher order derivatives of the Gaussian<br />

pdf can be taken to obtain Gaussian wavelets of various orders such as the seconde<br />

2<br />

derivative of the Gaussian pdf: the Mexican Hat wavelet ψ(t) =<br />

π 0.25√ 3 (t2 − 1)e −0.5t2<br />

which is a very popular wavelet for ECG processing [2, 20] too. When the question is<br />

posed what wavelet to use, the answer depends heavily on the application and signals at<br />

hand. Considering the large variety of functions that obey the admissability conditions<br />

there is a good chance that one can do better than by choosing a popular and well-known<br />

wavelet.<br />

3.2 Wavelets from filter banks<br />

The wavelet transform can also be performed on discrete time signals, thereby still using<br />

a continuous wavelet function ψ(t). One way to do this is with discrete filters from<br />

filter banks as in the discrete wavelet transform (DWT). In this thesis the framework of<br />

DWT wavelets from filter banks as described in for example [107] is used as a setting for<br />

wavelet design. This framework has a number of advantages and restrictions that make<br />

it a convenient setting for wavelet design. Filter banks may have, when imposed, five<br />

important properties of interest to us, which are further addressed below:<br />

• Perfect reconstruction<br />

• Orthogonality of the filter bank and the underlying wavelet based multi-resolution<br />

structure


28 CHAPTER 3. WAVELET TRANSFORMATIONS<br />

Analysis bank<br />

Synthesis bank<br />

x(z)<br />

H 0 (z)<br />

H 0 (z)x(z)<br />

2<br />

a(z)<br />

2<br />

½(H 0 (z)x(z)+<br />

H 0 (-z)x(-z))<br />

F 0 (z)<br />

H 1 (z)<br />

H 1 (z)z(z)<br />

2<br />

b(z)<br />

2<br />

½(H 1 (z)x(z)+<br />

H 1 (-z)x(-z))<br />

F 1 (z)<br />

+<br />

z -N x(z)<br />

Figure 3.2: Wavelet analysis and synthesis<br />

• Flatness of the filters and vanishing moments in the wavelets<br />

• Smoothness of the wavelets<br />

• Linear phase<br />

In discrete-time the input sequence is processed by two filters in parallel that split<br />

the input sequence in terms of frequency. These filters relate to the wavelet function<br />

in such a way that the downsampled output corresponds to the filtering of the signal<br />

with the wavelet function at a particular scale. The discrete-time input signal {x n } is<br />

essentially fed through a low-pass filter H 0 (z) and a high-pass filter H 1 (z) in parallel<br />

[38]. The filters H 0 (z) and H 1 (z) form a filter bank. The pair of filters in the filter bank<br />

are designed in such a way that after downsampling by a factor 2 no information is lost<br />

and there is no redundancy; they are critically sampled. In formulas, downsampling with<br />

a factor two will be written as ↓2. The output of the low-pass channel is a sequence of<br />

approximation coefficients (alias scaling coefficients) and the output from the high-pass<br />

channel a sequence of detail coefficients (alias wavelet coefficients). The original input<br />

can be reconstructed by upsampling and filtering each channel with the corresponding<br />

synthesis filter F 0 (z) or F 1 (z) and adding the two channels as illustrated in Figure 3.2.<br />

The downsampling can be implemented with frequency modulation [107]:<br />

a(z) =↓2H 0 (z)X(z) = 1 2<br />

(<br />

H 0 (z 1 2 )X(z<br />

1<br />

2 ) + H0 (−z 1 2 )X(−z<br />

1<br />

2 )<br />

)<br />

. (3.5)<br />

The aliasing term (with −z) cancels the odd part. This works similarly for the detail<br />

coefficients b(z). To upsample a(z) again take a(z 2 ), which comes down to inserting a 0<br />

after each element in the sequence.<br />

Example 3.2.1. As an example consider the Haar wavelet with low-pass filter H 0 (z) =<br />

1<br />

2 + 1 2 z and high-pass filter H 0(z) = − 1 2 + 1 2z. For input x = {4, 6, 8, 7} the low-pass<br />

output will be a = {5, 7.5} and the high-pass output b = {−1, 0.5}.


3.2. WAVELETS FROM FILTER BANKS 29<br />

3.2.1 Perfect reconstruction<br />

From this point on H 0 (z) and H 1 (z) are assumed to be FIR (finite impulse response)<br />

filters:<br />

H 0 (z) = c 0 + c 1 z −1 + . . . + c N z −N , (3.6)<br />

H 1 (z) = d 0 + d 1 z −1 + . . . + d N z −N . (3.7)<br />

This will result in wavelets that are compactly supported [107] and makes the conditions<br />

in this paragraph easy to satisfy.<br />

To have perfect reconstruction it means that if a signal is decomposed with an analysis<br />

filter bank and then reconstructed with a corresponding synthesis filter bank, the<br />

output signal is equal to the input signal expect for a possible delay. To ensure perfect<br />

reconstruction there may be no aliasing and no distortion.<br />

The aliasing effect [38] is present in both channels and must be canceled by the<br />

synthesis filters F 0 and F 1 :<br />

F 0 (z)H 0 (−z) + F 1 (z)H 1 (−z) = 0. (3.8)<br />

To ensure that there is no distortion in the output of Figure 3.2, the following condition<br />

must hold:<br />

F 0 (z)H 0 (z) + F 1 (z)H 1 (z) = 2z −Q , (3.9)<br />

where Q is the overall delay of the filter bank [107].<br />

The conditions (3.8) and (3.9) can be written in terms of a single modulation matrix:<br />

( ) ( )<br />

F0 (z) F 1 (z) H0 (z) H 0 (−z)<br />

=<br />

F 0 (−z) F 1 (−z) H 1 (z) H 1 (−z)<br />

3.2.2 Orthogonal filter banks<br />

( )<br />

2z<br />

−Q<br />

0<br />

0 2z −Q . (3.10)<br />

For orthogonal filter banks it holds that the impulse responses of the synthesis filters are<br />

the time reverses of the impulse responses of the analysis filters [106] as expressed by<br />

(3.13). To see this, we start by observing that the two filters in the filter bank H 0 (z)<br />

and H 1 (z) form a power complementary set [117, page 183]. Since on the unit circle<br />

the complex conjugate of z is z −1 the orthogonality condition from [117], taking the<br />

differences in notation into account, comes down to:<br />

H 0 (z −1 )H 0 (z) + H 1 (z −1 )H 1 (z) = c, (3.11)<br />

where c = 2 for consistency with (3.9), that is, to ensure conservation of energy.<br />

In this orthogonal framework the condition (3.8) can be enforced by constructing the<br />

synthesis filters from the analysis filters by alternating signs [31]:<br />

F 0 (z) = H 1 (−z) and F 1 (z) = −H 0 (−z). (3.12)


30 CHAPTER 3. WAVELET TRANSFORMATIONS<br />

To satisfy the condition in (3.9) the power complementary property can be used [117]:<br />

F 0 (z) = z −N H 0 (z −1 ) and F 1 (z) = z −N H 1 (z −1 ), (3.13)<br />

with the result that for orthogonal filters it holds that in (3.9) Q = N.<br />

To satisfy both (3.12) and (3.13) for an N th order filter one can construct the highpass<br />

filter from the low-pass filter by the alternating flip construction [31]. From the left<br />

side of (3.12) and (3.13) it is obtained that:<br />

After substituting z → −z it is obtained that:<br />

H 1 (−z) = z −N H 0 (z −1 ). (3.14)<br />

H 1 (z) = (−z) −N H 0 (−z −1 ). (3.15)<br />

And from the right side of (3.12) and (3.13) it is obtained that:<br />

After substituting z → z −1 it is obtained that:<br />

H 1 (z −1 ) = z N F 1 (z) = −z N H 0 (−z). (3.16)<br />

H 1 (z) = −z −N H 0 (−z −1 ). (3.17)<br />

From (3.15) and (3.17) it follows that N must be odd yielding the alternating flip construction:<br />

H 1 (z) = (−z) −N H 0 (−z −1 ), for N = 2n − 1. (3.18)<br />

In terms of (3.6) and (3.7) with N = 2n − 1 it follows that:<br />

d k = (−1) k c N−k (k = 0, 1, . . . , N). (3.19)<br />

Filters that are constructed in this manner are called quadrature mirror filters [38, 86, 30].<br />

For orthogonality the following constraints hold for the scaling filter coefficients c k<br />

and the wavelet filter coefficients d k of H 0 (z) and H 1 (z) respectively:<br />

Normalization ∑ N<br />

k=0 c2 k = 1 and ∑ N<br />

k=0 d2 k = 1<br />

Double-shift orthogonality ∑ N<br />

k=0 c kc k−2l = 0 and ∑ N<br />

k=0 d kd k−2l = 0 ∀l ∈ Z\ {0}<br />

Double-shift orthogonality between the filters ∑ N<br />

k=0 c kd k−2l = 0 ∀l ∈ Z<br />

In the last two conditions negatively indexed coefficients are all zero by convention (c 0 ,<br />

d k for k < 0 or k > N).


3.2. WAVELETS FROM FILTER BANKS 31<br />

Level j+1<br />

a (j-1)<br />

Level j<br />

H 0 2<br />

H 1 2<br />

a (j)<br />

b (j)<br />

H 0 2<br />

H 1 2<br />

a (j+1)<br />

b (j+1)<br />

Figure 3.3: Wavelet analysis tree structure<br />

3.2.3 Multi-resolution analysis<br />

The frequency localization for the discrete wavelet transform comes from using dilated<br />

basis functions, known as the scaling and wavelet function, as will be discussed in Section<br />

3.2.4. However this multi-resolution analysis can be implemented by a cascade of filter<br />

banks, consisting of an orthogonal pair of low- and high-pass filters. The output of the<br />

low-pass filter is used as the input of the next level as in Mallat’s algorithm [79]. This<br />

procedure is illustrated in Figure 3.3 and gives rise to multi-resolution analysis (MRA).<br />

Note that we now have approximation and detail coefficients at various levels j that<br />

correspond to dyadic scales 2 j .<br />

The basic principle is as follows: One starts with a signal x and feeds it through an<br />

analysis filter bank. The output of the low-pass filter is then a (1) =↓2H 0 x and that of<br />

the high-pass filter is b (1) =↓2H 1 x. Next a (1) is used as the input of the filter bank to<br />

yield a (2) =↓2H 0 a (1) and b (2) =↓2H 1 a (1) . Then again repeating this procedure a number<br />

of times, the output of the low-pass filter is used so that one eventually ends up with<br />

{b (1) , b (2) , . . . , b (L) , a (L) }. The signal x can be reconstructed by a cascade of synthesis<br />

filter banks. The length of the signal x and the size of the filter N determines how many<br />

levels this cascade can have. If the maximum number of levels L has been reached, the<br />

signal is said to be fully decomposed.<br />

3.2.4 Wavelet and scaling functions<br />

In this section it will be discussed how filter banks can be interpreted using the wavelet<br />

paradigm. For an excellent exposition on this topic the reader is referred to [107].<br />

Consider the following multi-resolution structure for the L 2 -space, consisting of a<br />

nested sequence of linear subspaces V l , called approximation spaces:<br />

. . . ⊂ V 1 ⊂ V 0 ⊂ V −1 ⊂ . . . , (3.20)<br />

with ⋂ l∈Z V l = {0} and ⋃ l∈Z V l = L 2 (R), where the bar indicates closure. Assume<br />

that there exists a function ϕ(t), called the scaling function, that spans the subspace V 0 .


32 CHAPTER 3. WAVELET TRANSFORMATIONS<br />

This scaling function generates a shift invariant orthonormal basis {ϕ(t − k)|k ∈ Z} of<br />

V 0 . From V 0 the orthonormal bases are induced for all spaces V l [80]:<br />

ϕ (l)<br />

k (t) = √ 1 ( t − 2 l )<br />

ϕ k (<br />

2<br />

l 2 l = 2 −l<br />

2 ϕ 2 −l t − k )<br />

V l = span{ϕ (l)<br />

k<br />

|k ∈ Z}. (3.21)<br />

Furthermore, consider the subspaces W l called detail spaces, that are the orthogonal<br />

complements of the approximation spaces V l :<br />

V l ⊕ W l = V l−1 . (3.22)<br />

Assume that the approximation space W 0 is spanned by a the wavelet function ψ(t), and<br />

its dilated versions are shift-invariant orthonormal bases for the spaces W l :<br />

ψ (l)<br />

k (t) = √ 1 ( t − 2 l )<br />

ψ k (<br />

2<br />

l 2 l = 2 −l<br />

2 ψ 2 −l t − k )<br />

W l = span{ψ (l)<br />

k<br />

|k ∈ Z}. (3.23)<br />

From (3.20) and (3.21), it follows that there exists a sequence of coefficients {c k } such<br />

that the scaling function can be written as a linear combination of dilated versions of<br />

itself in the dilation equation:<br />

ϕ(t) = √ 2<br />

N∑<br />

c k ϕ(2t − k). (3.24)<br />

k=0<br />

Likewise from (3.20), (3.22) and (3.23), it follows that there exists a sequence of<br />

coefficients {d k } such that the wavelet function can be written as a linear combination<br />

of dilated versions the scaling function in the wavelet equation:<br />

ψ(t) = √ 2<br />

N∑<br />

d k ϕ(2t − k). (3.25)<br />

k=0<br />

The coefficients c k and d k in (3.24) and (3.25) are the same coefficients as in Section<br />

3.2.2, with all properties involved, linking the filter coefficients to the wavelet and<br />

scaling function. This allows us to reverse the approach and given coefficients c k and<br />

d k try to find a scaling function ϕ(t) and a wavelet function ψ(t). An iteration scheme<br />

which allows for the exact computation of ϕ(t) at all the dyadic points up to an arbitrary<br />

resolution can be found, for instance, in [107, section 6.1]. We have adopted this<br />

method in this work. It is important to note that it may well happen that the resulting<br />

sequence exhibits a discontinuous and fractal structure and may not converge to an actual<br />

function, i.e., ϕ(t) and ψ(t) may not exist for a given set of filters. The continuity<br />

of the wavelet and scaling functions is further investigated in terms of the “joint spectral<br />

radius” in [110, Section 2.2].


3.2. WAVELETS FROM FILTER BANKS 33<br />

Now let f(t) be a function, corresponding to a sequence {x k } of wavelet coefficients,<br />

that can be expressed in terms of the wavelet basis for V 0 as:<br />

f(t) = ∑ k<br />

x k ϕ(t − k), (3.26)<br />

then the coefficients resulting from the DWT give an expansion for f(t) in terms of the<br />

orthonormal functions ϕ (l)<br />

k<br />

(t) and ψ(l)<br />

k<br />

(t):<br />

f(t) = ∑ k<br />

a (L)<br />

k<br />

ϕ(L) k<br />

L (t) + ∑ ∑<br />

where L is no more than the maximum level as in Section 3.2.3.<br />

wavelet structure it then also holds [91] that:<br />

a (l)<br />

k<br />

=<br />

b (l)<br />

k<br />

=<br />

∫ ∞<br />

−∞<br />

∫ ∞<br />

−∞<br />

l=1<br />

k<br />

b (l)<br />

k<br />

ψ(l) k<br />

(t), (3.27)<br />

For an orthogonal<br />

ϕ (l)<br />

k<br />

(t)f(t)dt, (3.28)<br />

ψ (l)<br />

k<br />

(t)f(t)dt, (3.29)<br />

which connects the continuous wavelet transform in (3.1) and discrete wavelet transform<br />

(3.29). Note that with the discrete wavelet transform these expansion coefficients {a (l)<br />

k }<br />

and {b (l)<br />

k<br />

} can be calculated without the explicit use of the scaling and wavelet function,<br />

but only by making use of a filter bank, using the approach in Section 3.2.3.<br />

3.2.5 Vanishing moments<br />

For some applications it is beneficial that the wavelets have vanishing moments. If a<br />

wavelet has p vanishing moments then the wavelet function ψ(t) is orthogonal to polynomials<br />

of up to degree p − 1 [80]. This leads to the following equation, which is familiar<br />

for moments from physics and statistics:<br />

∫ ∞<br />

−∞<br />

t k ψ(t)dt = 0, for 0 ≤ k < p. (3.30)<br />

Due to the orthogonality between the wavelet and scaling function it follows that if a<br />

wavelet has p vanishing moments then the space V 0 = span{ϕ k |k ∈ Z} spanned by<br />

the scaling function contains all the polynomials of degree less then p [107, Section<br />

7.1]. As familiar from Taylor series, smooth functions can locally be approximated by<br />

polynomials. If a wavelet has p vanishing moments, then the detail signal will contain no<br />

energy of polynomials up to degree p − 1. If a signal f(t) can locally in neighborhood v<br />

be represented as a polynomial part f v,p (t) with degree less then p and some error term<br />

ε v (t) then the wavelet transform of f will be [80, Section 6.1]:<br />

∫ ∞<br />

W fv (τ, σ) = f v (t) 1 ( ) ∫ t − τ<br />

∞<br />

√ ψ dt = ɛ v (t) 1 ( ) t − τ<br />

√ ψ dt = W ɛv (τ, σ).<br />

−∞ σ σ<br />

−∞ σ σ<br />

(3.31)


34 CHAPTER 3. WAVELET TRANSFORMATIONS<br />

When attempting to locally approximate the signal f with polynomials of degree less<br />

then p, ɛ represents the approximation error. The approximation error will have a nonzero<br />

contribution to the detail coefficients. This property can be used to measure the<br />

Lipschitz regularity (also known as the Hölder exponent) of a signal [80, Section 6.1].<br />

The approximation error decays with the level j (scale 2 j ) and the number of vanishing<br />

moments as O(2 jp ) [107, Section 7.1].<br />

For discrete time signals and wavelets, vanishing moments have to be imposed on the<br />

impulse response of H 1 (z), that is, moments up to order p − 1 have to be imposed on the<br />

wavelet filter. To have one vanishing moment amounts to the condition d 0 +d 1 +. . .+d N =<br />

0 which is equivalent to the commonly imposed condition that the integral of the mother<br />

wavelet ψ(t) is equal to zero: ∫ ∞<br />

ψ(t)dt = 0. In Section 3.2.4 it is discussed how the<br />

−∞<br />

wavelet function is obtained from the filter coefficients. For a wavelet filter to have p<br />

vanishing moments the constraints on the high-pass coefficients are:<br />

N∑<br />

l k d l = 0, for 0 ≤ k < p. (3.32)<br />

l=0<br />

The conditions on the wavelet filter (3.32) to have p vanishing moments can be translated<br />

to the scaling and wavelet function using the dilation equation (3.24) and the wavelet<br />

equation (3.25). From this it follows that the conditions on the filter are a linear combination<br />

of the conditions on the wavelet function (3.30) to have p vanishing moments,<br />

vice versa.<br />

For a wavelet filter to have vanishing moments, it is required that H 0 (z) and H 1 (z)<br />

have zeros at z = −1 and z = 1 respectively [80, Theorem 7.4][107, Theorem 7.1]. For<br />

filters this corresponds to requiring zeros at respectively π radians (Nyquist frequency)<br />

for the scaling filter and 0 radians for the wavelet filter. The zeros for the scaling filter at π<br />

radians make it a low-pass filter with a corresponding degree of flatness and the wavelet<br />

filter a high-pass filter with a certain degree of flatness. The Daubechies orthogonal<br />

wavelet family is constructed by using all freedom to impose vanishing moments.<br />

3.2.6 Linear phase<br />

As discussed on page page 22, linear phase helps to avoid phase distortion. There only<br />

exists a single orthogonal wavelet with linear phase: the Haar wavelet (see e.g. [120]).<br />

In order to obtain both orthogonality and linear phase simultaneously, the theory of<br />

multiwavelets (see Section 3.4) can be employed.<br />

3.2.7 Polyphase filtering<br />

In the traditional discrete wavelet transform approach,the high- and low-pass filters operate<br />

at full rate, after which the outputs are downsampled. In other words: half of the<br />

processed information is discarded. The use of polyphase filters avoids this, by first splitting<br />

the input into two or more (M) phases, then applying a polyphase filter to each phase<br />

separately and finally combining the polyphase outputs to generate the intended output.


3.2. WAVELETS FROM FILTER BANKS 35<br />

X(z)<br />

2<br />

X e<br />

(z)<br />

H 0,e<br />

(z)<br />

z -1<br />

2<br />

z -1 X o<br />

(z)<br />

H 0,o<br />

(z)<br />

+ (H 0 (z)X(z)) e<br />

Figure 3.4: Polyphase low-pass filter<br />

The polyphase filters operate at reduced rate 1 M<br />

. Suppose an input signal X(z) and an<br />

analysis low-pass filter H 0 (z). The goal is to obtain ↓2H 0 X; the downsampled result of<br />

filtering X(z) with H 0 . To implement this with polyphase filters, with e.g., M = 2, take<br />

H 0,e (z) = c 0 +c 2 z −1 +. . .+c 2n−2 z −(n−1) and H 0,o (z) = c 1 +c 3 z −1 +. . .+c 2n−1 z −(n−1) as<br />

respectively the even and odd phase of the filter H 0 (z), and X e (z) and X o (z) as respectively<br />

the even and odd phase of the input signal X(z). (Since M = 2 it is convenient to<br />

call one of the phases “even” and the other one “odd”.) Then the desired output ↓2H 0 x<br />

can be calculated [107, Section 4.2] with polyphase filters as:<br />

Z {↓2H 0 x} = (H 0 (z)X(z)) e<br />

= X e (z)H 0,e (z) + z −1 H 0,o (z)X o (z)<br />

= [ H 0,e (z) H 0,o (z) ] [ ]<br />

X e (z)<br />

z −1 ,<br />

X o (z)<br />

(3.33)<br />

This process is illustrated in Figure 3.4. The basic idea behind this approach to find<br />

the even part, magister dixit, “So it is really odd plus odd, and even plus even” [107, p.<br />

115].<br />

Now consider a filter bank with low-pass filter H 0 and high-pass filter H 1 . These<br />

filters can be split into two phases and a polyphase matrix H p (z) can be constructed as:<br />

[ ]<br />

H0,e (z) H<br />

H p (z) =<br />

0,o (z)<br />

, (3.34)<br />

H 1,e (z) H 1,o (z)<br />

Then this polyphase matrix H p (z) can be used to implement the low- and high-pass filter<br />

in parallel by acting on the same polyphase input:<br />

[ ] [ ]<br />

V0 (z) (H0 (z)X(z))<br />

=<br />

e<br />

=<br />

V 1 (z) (H 1 (z)X(z)) e<br />

[ ] [ ]<br />

H0,e (z) H 0,o (z) Xe (z)<br />

H 1,e (z) H 1,o (z) z −1 . (3.35)<br />

X o (z)<br />

For reconstruction the synthesis filter can be implemented as a polyphase filter F p (z)


36 CHAPTER 3. WAVELET TRANSFORMATIONS<br />

X(z)<br />

2<br />

X e (z)<br />

V 0 (z)<br />

H<br />

z -1 p (z) F p (z)<br />

X o (z)<br />

V 1 (z)<br />

2<br />

F 0,o (z)V 0 (z)<br />

+F 1,o V 1 (z)<br />

F 0,e (z)V 0 (z)<br />

+F 1,e V 1 (z)<br />

2<br />

2 +<br />

Z -1<br />

^<br />

X(z)<br />

Figure 3.5: Polyphase analysis and synthesis filters<br />

too:<br />

F p (z) =<br />

[<br />

F0,o (z)<br />

]<br />

F 1,o (z)<br />

F 0,e (z) F 1,e (z)<br />

ˆX(z) = [ z −1 1 ] F p (z 2 )<br />

[<br />

V0 (z)<br />

V 1 (z)<br />

(3.36)<br />

]<br />

. (3.37)<br />

In the polyphase approach the analysis and synthesis filters are adjacent, in contrast to<br />

the filterbank approach where they are separated by down- and upsampling. As a result<br />

the condition for perfect reconstruction becomes elegant and simple for polyphase filters:<br />

F p (z)H p (z) = z −q I. (3.38)<br />

For practical purposes, the delay q should be as small as possible. The overall delay<br />

induced by the filter system due to the fact that the filters are causal is Q = 2q + 1,<br />

so that ˆX(z) = z −Q X(z). In the case of orthogonality as in Section 3.2.2 it holds<br />

that Q = N = 2n − 1. The elegant condition in (3.38) facilitates the parameterization<br />

of wavelet filters. Due to the fact that the polyphase filters operate at half rate this<br />

approach is also beneficial from a computational viewpoint.<br />

3.3 The stationary wavelet transform<br />

The multi-resolution approach discussed in Section 3.2.3 is well suited for compression<br />

purposes since it is critically sampled, i.e., all resulting wavelet coefficients are required<br />

to ensure perfect reconstruction without redundancy. If a sequence x of length 2 m is<br />

filtered with wavelet filters H 0 and H 1 , then x is mapped to ↓2H 0 x and ↓2H 1 x by means<br />

of a projection. Both ↓2H 0 x and ↓2H 1 x will then have a length of approximately 2 m−1 ,<br />

effectively reducing the resolution by a factor two. After rewriting the upper right-hand<br />

side of (3.23) as<br />

ψ (l)<br />

−l<br />

k<br />

(t) = 2 2 ψ(2 −l (t − 2 l k))<br />

= 2 −l<br />

2 ψ(2 −l (t − τ)), τ = 2 l k, k ∈ Z, (3.39)<br />

one can see that the timeshifts of the dilated wavelet functions 2 −l<br />

2 ψ(2 −l t) are integer<br />

multiples of 2 l with an initial offset of zero.


3.3. THE STATIONARY WAVELET TRANSFORM 37<br />

When employing the wavelet transform for detection purposes, this approach has two<br />

important drawbacks, in casu the lack of shift invariance and the loss of resolution at<br />

coarse scales. These problems can be avoided by using an overcomplete wavelet transform.<br />

The so-called stationary wavelet transform [91] achieves this by taking timeshifts equal<br />

to all integer values instead of 2 l as in (3.39).<br />

From (3.29) the connection between the continuous and discrete wavelet transform<br />

becomes clear. If (3.29) is applied at an appropriate resolution at each scale, one obtains<br />

a transform with the same resolution as the original signal x of length n = 2 m . The<br />

regular DWT uses downsampling at each successive scale, which causes loss of resolution<br />

at those scales. One way to avoid this is to perform the DWT 2 L times (L being the<br />

number of scales in the MRA) and each time shifting the sequence x by one.<br />

The results can then be combined to have full resolution at each scale; all the shifted<br />

wavelet transforms combined give a shift invariant, redundant transform. A major drawback<br />

of this approach is that it is very inefficient, since such a large number of shifts are<br />

not required at fine scales.<br />

Another approach is to split the wavelet decomposition at each scale. Let a (l−1) be<br />

the input to scale l. Then one can perform a wavelet transform on the approximation<br />

signal at the preceding level a (l−1) and on its time-shifted version z −1 a (l−1) . This results<br />

in a tree that branches out exponentially fast compared to the the decomposition tree in<br />

the regular discrete wavelet transform<br />

a (l),0 =↓2H 0 a (l−1)<br />

b (l),0 =↓2H 1 a (l−1)<br />

a (l),1 =↓2H 0 z −1 a (l−1)<br />

b (l),1 =↓2H 1 z −1 a (l−1) .<br />

(3.40)<br />

This process is also illustrated in Figure 3.6. Due to recursion we get<br />

a (l),[0,S] =↓2H 0 a (l−1),S , (3.41)<br />

where S is a bit string that encodes through what downsampling types (even or odd)<br />

the signal has passed, with the most recent type up front. A 0 in S indicates that the<br />

input passed through a downsampling ↓ 2, i.e., the regular detail and approximation<br />

coefficients are involved, and an 1 indicates that it has passed through a delay followed<br />

by downsampling ↓ 2z −1 , i.e., the detail and approximation coefficients of the shifted<br />

signal are involved.<br />

At scale k (scale k = 0 corresponds to the signal) there are thus 2 k sets of approximation<br />

and detail coefficients, with each a associated bit string. The sets of detail<br />

coefficients at each scale have to be interlaced such that a single set of coefficients is


38 CHAPTER 3. WAVELET TRANSFORMATIONS<br />

a (l-1),s<br />

a (l),s<br />

H 0 2<br />

a (l),[0,s]<br />

H 0<br />

a (l+1),[0,s]<br />

H 1<br />

b (l),s<br />

2<br />

z -1 2<br />

z -1 2<br />

a (l),[1,s]<br />

H 1<br />

H 0<br />

b (l+1),[0,s]<br />

a (l+1),[1,s]<br />

2<br />

z -1 2<br />

H 1<br />

b (l+1),[1,s]<br />

a (l+1),[0,0,s]<br />

a (l+1),[1,0,s]<br />

a (l+1),[0,1,s]<br />

a (l+1),[1,1,s]<br />

Figure 3.6: Stationary wavelet transform<br />

acquired at full resolution. For example at level l + 1 we have:<br />

b even (l+1),[0,S] = b (l+1),[0,0,S] , (3.42)<br />

b (l+1),[0,S]<br />

odd<br />

= b (l+1),[1,0,S] , (3.43)<br />

b even (l+1),[1,S] = b (l+1),[0,1,S] , (3.44)<br />

b (l+1),[1,S]<br />

odd<br />

= b (l+1),[1,1,S] , (3.45)<br />

b (l+1),S<br />

even = b (l+1),[0,S] , (3.46)<br />

b (l+1),S<br />

odd<br />

= b (l+1),[1,S] . (3.47)<br />

This process is displayed in Figure 3.7. For the maximum level L the sets of approximation<br />

coefficients also have to be interlaced, likewise to obtain a wavelet decomposition<br />

with full resolution at each dyadic scale.<br />

The stationary wavelet transform can also be calculated in polyphase. The stationary<br />

wavelet transform in polyphase representation is partially displayed in Figure 3.8, where<br />

H p is a polyphase filter with two inputs (the phases) and two outputs.<br />

Another approach to calculate the stationary wavelet transform is to modify the<br />

filters. Such an approach, which gives the same results, is discussed in [91].


3.3. THE STATIONARY WAVELET TRANSFORM 39<br />

b (l+1),[0,0,s]<br />

b (l+1),[1,1,s] 2 z<br />

2<br />

b (l+1),[1,0,s]<br />

2 z<br />

b (l+1),[0,1,s]<br />

2<br />

+<br />

+<br />

b (l+1),[0,s]<br />

b (l+1),[1,s]<br />

2<br />

2 z +<br />

b (l+1),s<br />

Figure 3.7: Interlacing for the stationary wavelet transform<br />

a (l-1),s<br />

2<br />

a (l),[0,s]<br />

z -1 2<br />

H p<br />

b (l),[0,s]<br />

2<br />

a (l),[1,s]<br />

z -1 H p<br />

b (l),[1,s]<br />

2 z +<br />

b (l),s<br />

Figure 3.8: Stationary wavelet transform in polyphase representation


40 CHAPTER 3. WAVELET TRANSFORMATIONS<br />

3.4 Multiwavelets<br />

Regular scalar wavelets have a number of limitations. For example both linear phase and<br />

orthogonality is only possible for the Haar wavelet as discussed in Section 3.2.6. Another<br />

problem is that for a given choice of basis function this basis function may correlate with<br />

the features in a signal is such manner, that it is impossible to distinguish these features<br />

in the wavelet domain. Multiwavelets are a generalization of wavelets in the sense that<br />

instead of that V 0 is spanned in the L 2 space by a basis generated (through integer shifts)<br />

from a single scalar function ϕ(t), it is spanned by a vector (multiscaling) function ϕ(t)<br />

[<br />

T<br />

[119, 76, 77, 107], which is a vector of r scaling functions ϕ(t) = ϕ [0]<br />

(t), . . . , ϕ [r−1]<br />

(t)]<br />

.<br />

The dilation equation is a vector generalization of (3.24):<br />

ϕ(t) = √ 2<br />

N∑<br />

C k ϕ(2t − k), (3.48)<br />

k=0<br />

with H 0 (z) = C 0 + C 1 z −1 + . . . + C N z −N where each C k is an r × r coefficient matrix.<br />

Likewise a multiwavelet function, which is a vector function ψ(t) is introduced as:<br />

ψ(t) = √ 2<br />

N∑<br />

D k ϕ(2t − k). (3.49)<br />

k=0<br />

The entries of ψ(t) (with their integer translates) constitute a basis for W 0 with the multiwavelet<br />

filter H 1 (z) = D 0 + D 1 z −1 + . . . + D N z −N . The Smith-Barnwell orthogonality<br />

conditions can be imposed as:<br />

H 0 (z)H 0 (z −1 ) T + H 0 (−z)H 0 (−z −1 ) T = 2I r , (3.50)<br />

H 1 (z)H 1 (z −1 ) T + H 1 (−z)H 1 (−z −1 ) T = 2I r , (3.51)<br />

H 0 (z)H 1 (z −1 ) T + H 0 (−z)H 1 (−z −1 ) T = 0, (3.52)<br />

generalizing (3.10). The vectorfunctions ϕ(t) and ψ(t) both have compact support in<br />

the interval [0, N] due to the assumed FIR property of the related filters. They generate<br />

a multiresolution structure for the inner-product space L 2 (R): . . . , V −1 , V 0 , V 1 , . . ., with<br />

⋂<br />

l∈Z V l = {0} and ⋃ k∈Z V k = L 2 (R), analogous to the regular DWT case. Note however<br />

that all inputs and outputs are vector sequences with r components. The input is now<br />

recomposed of the vector sequence containing the r phases.<br />

Multiwavelets can also be implemented with polyphase filters. The first step is to<br />

split the input signal X(z), the low-pass filters H 0 (z) and the high-pass filters H 1 (z) into<br />

2r phases.<br />

X(z) = X e (z 2 ) + z −1 X o (z 2 ) (3.53)<br />

H 0 (z) = H 0,e (z 2 ) + z −1 H 0,o (z 2 ) (3.54)<br />

H 1 (z) = H 1,e (z 2 ) + z −1 H 1,o (z 2 ). (3.55)


3.4. MULTIWAVELETS 41<br />

6<br />

6<br />

a (2,0)<br />

z -1 x<br />

a (1,0)<br />

2<br />

z -1 b (1,2) b (2,2)<br />

z -1<br />

a (1,1) z -1<br />

2<br />

a (2,1)<br />

H p<br />

(z)<br />

b z<br />

6<br />

2<br />

H p<br />

(z)<br />

b (2,0)<br />

z -1 6<br />

a (1,2)<br />

2<br />

a (2,2)<br />

z -1 6<br />

b (1,1)<br />

2<br />

b (2,1)<br />

6<br />

z -1<br />

2<br />

Figure 3.9: Multiwavelet with r = 3 in polyphase representation<br />

Next one constructs the 2r × 2r polyphase FIR filter H p (z):<br />

( )<br />

H0,e (z) H<br />

H p (z) =<br />

0,o (z)<br />

. (3.56)<br />

H 1,e (z) H 1,o (z)<br />

Then the polyphase filter H p (z) is applied, resulting in r vectors with detail coefficients<br />

b (l,m)<br />

k<br />

(l being the scale and m = 0, . . . , r − 1) and r vectors with approximation coefficients<br />

a (l,m)<br />

k<br />

. The approximation coefficients are then split into two phases and iterated<br />

through the polyphase filter H p (z) as illustrated in Figure 3.9. We have:<br />

Y k (z) =<br />

2r−1<br />

∑<br />

m=0<br />

H km (z)U m (z), (3.57)<br />

where H km (z) is the transfer function from the m th input to the k th output. And this<br />

can be rewritten in vector form as Y (z) = H p (z)U(z). The orthogonality conditions<br />

translate in terms of the polyphase representation into:<br />

H p (z)H p (z −1 ) T = I 2r , (3.58)<br />

where I 2r is the 2r × 2r identity matrix.<br />

Lossless systems [116, 117] or stable all-pass systems are systems that retain the<br />

energy from the (Fourier transformable) input U to the output Y (with a possible scale


42 CHAPTER 3. WAVELET TRANSFORMATIONS<br />

factor):<br />

1<br />

2π<br />

∫ 2π<br />

0<br />

|Y (e iω )| 2 dω = c<br />

2π<br />

∫ 2π<br />

0<br />

|U(e iω )| 2 dω, c ∈ R + . (3.59)<br />

It holds [117] for the transfer matrix of a stable all-pass system that:<br />

H(z) † H(z) = cI, ∀|z| = 1, c ∈ R + , (3.60)<br />

where † indicates the Hermitian transpose or conjugate transpose which is defined as<br />

transposition followed by complex conjugation: (A † ) a,b = A b,a . In order to build in the<br />

condition (3.58) the constant c is chosen to be c = 1. Note that wavelets filters are clearly<br />

stable since they are FIR and that they are all-pass. As a result it additionally holds<br />

that:<br />

H(z) † H(z) ≤ I, |z| > 1, (3.61)<br />

H(z) † H(z) ≥ I, |z| < 1. (3.62)<br />

Similarly from the properties of lossless systems in [117]:<br />

˜H(z)H(z) = I s , z = e iω , (3.63)<br />

˜H(z) = H(z −1 ) T ∗ , (3.64)<br />

where the subscript H(z) ∗ stands for conjugation of the coefficients of H(z), i.e., H(z)<br />

has to be unitary (orthogonal for real matrices) on |z| = 1. Since we have real coefficients<br />

in H p (z) in (3.58) and because (3.64) extends to the whole complex plane due to analytic<br />

continuation, (3.58) is equivalent to the property of lossless systems in (3.64). Hence the<br />

condition of orthonormality for multiwavelets comes down to requiring that H p (z) is<br />

lossless. A real lossless polyphase matrix H p (z) of order n − 1 thus corresponds with a<br />

pair of polynomial matrices H 0 (z) and H 1 (z) that form an orthogonal FIR (multi)wavelet<br />

filterbank of order 2n − 1. An interpolation condition on the unit circle for the lossless<br />

system can be used to impose a vanishing moment for the wavelets as will be further<br />

discussed in Section 5.3. This is useful for parameterization purposes.


Chapter 4<br />

Continuous-time analog implementation<br />

of wavelets<br />

In implantable medical devices such as pacemakers power consumption is a critical issue<br />

because battery lifetime is limited. This especially holds for sensing circuits since they<br />

are permanently active. To perform discrete-time digital signal processing, an analog<br />

to digital (A/D) converter and sampling (continuous time to discrete time) is required<br />

in order to transfer continuous-time analog sensor information to the digital domain.<br />

Depending on the number of bits used, the A/D conversion is a heavy power consuming<br />

operation. For power consumption considerations it therefore is preferable to perform as<br />

many computations as possible in the continuous-time analog domain. In earlier work<br />

[53, 50, 49, 54, 48] analog dynamic translinear systems (see Section 4.1) were used as a<br />

platform to implement continuous wavelet transforms (see Section 3.1). This involved the<br />

approximation of the wavelet transform with the impulse response of a linear system as<br />

discussed in Section 4.2. Using the technique of Padé approximation the authors of [53]<br />

obtained a rational approximation of the Laplace transform of the wavelet function under<br />

approximation. In Section 4.3 this approach, that only can approximate a limited number<br />

of wavelet functions, is discussed in more detail. A key problem of this Padé approach is<br />

that it does not behave equally well for each time instance. An new alternative approach,<br />

based on L 2 approximation, that can be used for a wider range of wavelet functions is<br />

discussed in Sections 4.4 and 4.5. This approach can even be employed to approximate the<br />

wavelet function of discrete wavelets such as the Daubechies 3 wavelet as demonstrated<br />

in Section 4.5.<br />

4.1 Dynamic translinear systems<br />

The value of an input or output signal is commonly represented in circuits in the voltage<br />

domain (volts). However Dynamic Translinear (DTL) circuits [90, 37] are current-based<br />

43


44 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

Linear log-domain filter<br />

Nonlinear integrator<br />

Linear integrator<br />

I in<br />

Log<br />

(Q 1 )<br />

V in<br />

+<br />

-<br />

Exp<br />

(Q 2 )<br />

I<br />

1/S<br />

(C)<br />

V c<br />

Exp<br />

(Q 4 )<br />

I out<br />

I in<br />

Q 1 Q 2 Q 3 Q 4<br />

C<br />

I0 2I0<br />

I out<br />

Figure 4.1: Linear log-domain lossy integrator. Note that Q 3 is not fundamental for the<br />

operation.<br />

circuits (amps). As a result capacitors and transistors are used in the DTL circuit,<br />

but no resistors. The voltage-to-current transformation corresponds to an exponential<br />

transformation [90] if, for example, bipolar transistors are used or CMOS transistors in<br />

the region where the Voltage density is small, i.e., in the weak inversion region.<br />

In Figure 4.1 it is illustrated how an integrator can be implemented in a DTL circuit.<br />

It is commonly said that the integrator is implemented int the “log-domain”, referring<br />

to the logarithmic relation between the internal state variables and capacitance voltages<br />

and input/output voltages, since the current-to-voltage transformation corresponds to a<br />

logarithmic transform and the voltage-to-current transformation to a exponential transform.<br />

The nonlinear integrator is the log-domain filter in this figure [37]. The outputs<br />

are related linearly to currents. The integrator block is nonlinear, however the overall<br />

scheme is linearized by transforming the input/output voltages to overall input/output<br />

currents respectively. It is well known that the for the product of two exponentials it<br />

holds that exp a exp b = exp a + b and therefore this logarithmic relationship in DTL


4.2. WAVELET TRANSFORMATIONS AS LINEAR SYSTEMS 45<br />

circuits makes it possible to implement multipliers. Furthermore the derivative of an<br />

exponential function is equal to the exponential function times the derivative of the exponent.<br />

Not only linear systems can be implemented with the DTL approach but also<br />

some non-linear operations, such as nonlinear differential equations. Four advantages of<br />

DTL circuits of interest are:<br />

1. The current based scheme is potentially less power-consuming than a voltage based<br />

scheme.<br />

2. The absence of resistors in the physical implementation allows a smaller circuit.<br />

3. Both linear systems as various non-linear operations (such as an RMS D/C converter,<br />

an oscillator with limit cycle or multipliers) can be implemented with DTL<br />

circuits.<br />

4. An attractive potential of the DTL technique are the dynamic possibilities: The<br />

behavior of the circuit can be changed almost instantaneously by changing the<br />

magnitude of the currents that determine the implemented state-space system.<br />

This can for example be used to scale the system such that a wavelet can be dilated<br />

in order to adapt to the specific frequency that is relevant for a sense amplifier in<br />

a given state.<br />

In [54] a DTL approach is described which aims to implement the Gaussian wavelet<br />

transform. This Gaussian wavelet transform is of interest for ECG processing (see e.g.<br />

[5, 102]). The implementation is done in the analog domain for the purpose of cardiac<br />

signal analysis. In Figure 4.2 a DTL implementation of a state-space system is displayed.<br />

The performance of such an implementation depends largely on the accuracy of the<br />

approximations involved in this approach. From a technological point of view, the quality<br />

of the hardware components used in the manufacturing process may have a considerable<br />

impact on the performance of the IC, but such issues will not be discussed here. From<br />

a conceptual point of view, one of the critical steps concerns the approximation of the<br />

Laplace transform of the (time-reversed and shifted) Gaussian wavelet function by means<br />

of a strictly proper rational function of low order. For this purpose the classical technique<br />

of Padé approximation [10, 19] was previously proposed. We will be discussing the<br />

drawbacks of this approach and propose an alternative approach: L 2 approximation of<br />

wavelets, that overcomes the problems discussed for the Padé approximation.<br />

4.2 Wavelet transformations as linear systems<br />

The available IC design methods only allow for a limited class (e.g. finite order and<br />

causal) of linear filters to be implemented. The IC design of linear filters (Section 2.8)<br />

of finite order is quite well understood. If a time signal f(t) is passed through a linear<br />

system, then f(t) is convoluted with the impulse response h(t) of that linear system,<br />

producing the output signal as in (2.26). If the continuous wavelet transform W (τ, σ)<br />

of f(t) associated with a given mother wavelet ψ(t) on a scale σ is considered, this


46 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

I A<br />

Integration section:<br />

Amatrix<br />

Vdc<br />

+<br />

A12<br />

-<br />

Wij<br />

I<br />

Aij (1nA<br />

<br />

L 1nA<br />

ij<br />

+<br />

Vdc A21<br />

-<br />

Vdc<br />

+<br />

A23<br />

-<br />

)<br />

I A12<br />

+<br />

Vdc A32<br />

-<br />

Vdc<br />

+<br />

A34<br />

-<br />

Current matrix I Aij<br />

(PMOS current mirros)<br />

I A21<br />

+<br />

Vdc A43<br />

-<br />

Vdc<br />

+<br />

A45<br />

-<br />

Input section: B vector<br />

+<br />

Vdc<br />

I in<br />

+<br />

Vdc<br />

- B10 0 -<br />

+<br />

Vdc A54<br />

-<br />

+<br />

Vdc A65<br />

-<br />

Vdc<br />

+<br />

-<br />

A56<br />

+<br />

Vdc A76<br />

-<br />

Vdc<br />

+<br />

A67<br />

-<br />

+<br />

Vdc<br />

-<br />

A87<br />

Vdc<br />

+<br />

A78<br />

-<br />

+<br />

Vdc A98<br />

-<br />

Vdc<br />

+<br />

-<br />

A89<br />

+<br />

Vdc A109 -<br />

I A910 I A1010<br />

Vdc<br />

+ +<br />

Vdc<br />

A910<br />

- -<br />

C 1 C 2 C 3 C 4 C 5 C 6 C 7<br />

I B<br />

A1010<br />

+<br />

C1<br />

-<br />

+<br />

-C2<br />

+<br />

C3<br />

-<br />

+<br />

C4<br />

-<br />

+<br />

-C5<br />

+<br />

C6<br />

-<br />

+<br />

-C7<br />

+<br />

C8<br />

-<br />

+<br />

C9<br />

-<br />

C 8 C 9 C 10<br />

W<br />

j<br />

I<br />

Cj (1nA<br />

)<br />

<br />

L 1nA<br />

j<br />

I C1<br />

I C2<br />

I C10<br />

+<br />

Vdc<br />

-<br />

I out<br />

Summation section:<br />

C vector<br />

I C<br />

Current matrix I Cj<br />

( PMOS current mirros)<br />

Figure 4.2: Implementation of a state-space system with a DTL circuit. Picture at the<br />

courtesy of Sandro Haddad. Note that in this specific case I C10 is negligible small [47,<br />

p. 181].<br />

transform is obtained as the integral defined in (3.1). Note that the wavelet transform<br />

at a fixed scale σ involves a linear filter operation. Therefore, the analog computation<br />

of W (τ, σ) can be achieved through the implementation of a linear filter of which the<br />

impulse response satisfies<br />

h(t) = 1 √ σ<br />

ψ<br />

( ) −t<br />

. (4.1)<br />

σ<br />

For linear systems of a finite order this equation can in general not be satisfied exactly,<br />

however a reasonably good approximation may be sufficient for the intended application.<br />

Equation (4.1) can also be reformulated in the Laplace domain:<br />

H(s) =<br />

∫ ∞<br />

0<br />

1<br />

√ σ<br />

ψ<br />

( ) −t<br />

e −st dt. (4.2)<br />

σ<br />

For obvious physical reasons only the hardware implementation of strictly causal stable<br />

filters of sufficiently low order is feasible. In other words, an implementable linear filter<br />

will have a (strictly) proper rational transfer function H(s) that has all its poles in the<br />

left half of the complex plane to ensure stability. In a causal system the degree of the<br />

numerator is less than or equal to the degree of the denominator, yielding a direct feed-


4.2. WAVELET TRANSFORMATIONS AS LINEAR SYSTEMS 47<br />

order norm shift 2.0 shift 2.5 shift 3.0 shift 3.5<br />

3 l 1 -norm 0.997 1.466 1.871 1.980<br />

5 l 1 -norm 0.117 0.265 0.496 0.789<br />

7 l 1 -norm 0.017 0.027 0.071 0.153<br />

9 l 1 -norm 0.016 0.002 0.007 0.021<br />

3 l 2 -norm 0.397 0.551 0.678 0.696<br />

5 l 2 -norm 0.048 0.101 0.178 0.269<br />

7 l 2 -norm 0.007 0.010 0.025 0.053<br />

9 l 2 -norm 0.010 0.001 0.003 0.007<br />

3 l ∞ -norm 0.490 0.539 0.485 0.430<br />

5 l ∞ -norm 0.077 0.151 0.232 0.303<br />

7 l ∞ -norm 0.005 0.017 0.041 0.076<br />

9 l ∞ -norm 0.002 0.001 0.004 0.011<br />

3 CPU time 0.630 0.820 1.380 1.700<br />

5 CPU time 0.860 0.990 1.240 1.430<br />

7 CPU time 17.010 2.340 1.980 1.820<br />

9 CPU time 933.170 62.390 10.390 5.520<br />

energy loss 5.5 · 10 −4 7.4 · 10 −6 3.5 · 10 −8 6.1 · 10 −11<br />

Table 4.1: Effect of time-shift on required order, norm of misfit and energy loss illustrated<br />

on the Gaussian wavelet. The approximations were determined by using a deterministic<br />

starting point as in Section 4.4.4. Note that the l 2 norm of the 9 th order approximation<br />

with a time-shift of two is less well than for the 7 th order approximation. This is due to<br />

the fact that in this case the deterministic starting point leads to a local optimum. If the<br />

7 th order approximation was used as a starting point for the 9 th order approximation a<br />

better result would have been obtained.<br />

through in the case of an equality. Strict causality ensures that the direct feed-through<br />

is absent.<br />

Due to this strict causality h(t) will be zero for negative t, so that any time-reversed<br />

mother wavelet ψ(−t) which does not have this property must be time-shifted, by some<br />

value t 0 , yielding, to facilitate an accurate approximation of its (correspondingly timeshifted)<br />

wavelet transform ˜W (τ, σ):<br />

where<br />

˜W (τ, σ) =<br />

∫ ∞<br />

−∞<br />

x(t) ˜ψ trunc (τ − t)dt, (4.3)<br />

˜ψ(t) = ψ(t 0 − t), (4.4)<br />

˜ψ trunc (t) =<br />

{<br />

˜ψ for t ≥ 0<br />

0 for t < 0<br />

(4.5)<br />

In case ψ(t 0 −t) is nonzero for t > t 0 , a truncation error results, which should be kept<br />

small. This is illustrated in Table 4.1, where the time-reversed and time-shifted Gaussian<br />

wavelet is approximated, using various delays. Note that an approximation error will also


48 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

occur due to the fact that a wavelet does not usually possess a rational Laplace transform<br />

which is a requirement that follows from the restrictions to linear filters of finite order.<br />

A possible approach for this filter synthesis problem is to use N building blocks with<br />

each an impulse response function h k , k = 1, . . . , N and to combine them with a weight<br />

vector w k such that h(t) = ∑ N<br />

k=1 w kh k (t) ≈ ˜ψ trunc (t). This approach is for example<br />

discussed in [12]. Due to the high requirements on power consumption this approach<br />

however is not well suited for the application at hand. The system should be tailored for<br />

the desired impulse response to keep the order and thus the power consumption as low<br />

as possible.<br />

4.3 Padé approximation of wavelet functions<br />

Padé approximation provides a method to obtain a rational approximation to a given<br />

function. In the current application a rational function in the Laplace domain is needed.<br />

Therefore the intended application requires one to work in the frequency domain. Any<br />

Padé approximation H(s) of ˜Ψ(s)<br />

{ }<br />

= L ˜ψ(t) is characterized by the property that the<br />

coefficients of the Taylor series expansion:<br />

˜Ψ(s) =<br />

∞∑<br />

k=0<br />

1<br />

k! ˜Ψ (k) (s 0 )(s − s 0 ) k (4.6)<br />

of ˜Ψ(s) around a selected point s = s 0 coincide with the corresponding Taylor series<br />

coefficients of H(s) up to the highest possible order, given the pre-specified degrees of the<br />

numerator and denominator polynomials of H(s). If we denote the Padé approximation<br />

H(s) at s = s 0 and of order (n, m) with n ≤ m by<br />

H(s) = p 0(s − s 0 ) n + p 1 (s − s 0 ) n−1 + . . . + p n<br />

(s − s 0 ) m + q 1 (s − s 0 ) m−1 + . . . + q m<br />

(4.7)<br />

then there are m+n+1 degrees of freedom, which generically makes it possible to match<br />

exactly the first m + n + 1 coefficients of the Taylor series expansion of ˜Ψ(s) around<br />

s = s 0 . In fact, this matching problem can easily be rewritten as a system of m + n + 1<br />

linear equations in the m + n + 1 variables p 0 , p 1 , . . . , p n , q 1 , . . . , q m . See, e.g., [10].<br />

This brings us to one of the main advantages of Padé approximation: the linear system<br />

of equations will generically yield a unique solution which is easy to compute. Moreover,<br />

a good match is guaranteed between the given function ˜Ψ(s) and its approximation H(s)<br />

in a neighborhood of the selected point s 0 . However, there are also some disadvantages<br />

which limit the practical applicability of this technique in the setting of this research.<br />

One important issue concerns the selection of the point s 0 . Note that a good approximation<br />

of ˜Ψ(s) over the entire (complex) Laplace domain is not a requirement per se.<br />

Instead, an approximation is needed which performs well when used for convolution in<br />

the time domain. Since the function ˜ψ(t) is a wavelet, it effectively will have compact<br />

support and in particular it should be approximated well in the region of the time domain<br />

where it ‘lives’: which is somewhere near t = 0. Now, the initial value theorem for the


4.3. PADÉ APPROXIMATION OF WAVELET FUNCTIONS 49<br />

0.8<br />

0.6<br />

0.4<br />

0.2<br />

0<br />

−0.2<br />

−0.4<br />

−0.6<br />

time-reversed/shifted gaussian wavelet<br />

5th order pade s 0 =0<br />

7th order Pade, s 0 →∞<br />

9th order generalization Pade<br />

−0.8<br />

0 1 2 3 4 5 6<br />

time−reversed/shifted gaussian wavelet<br />

10 4 5th order pade s 0<br />

=0<br />

7th order Pade, s →∞ 0<br />

10 3<br />

9th order generalization Pade<br />

10 2<br />

10 1<br />

10 0<br />

0 1 2 3 4 5 6<br />

Figure 4.3: Top figure: Padé approximations of the Gaussian wavelet. Lower figure: The<br />

value of all function has been increased by 1 such they are all positive and they can be<br />

plotted on a logarithmic scale. It is clear that the Padé approximation with s 0 → ∞ is<br />

unstable and the Padé approximation with s 0 = 0 has a poor fit at the beginning of the<br />

impulse response.


50 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

Laplace transformation (2.18) motivates the choice s 0 = ∞. This choice will lead to a<br />

good approximation of ˜ψ(t) near t = 0, as is demonstrated in Figure 4.3 for a 7th order<br />

approximation of the Gaussian wavelet.<br />

A second important issue concerns stability. The approximation h(t) of ˜ψ(t) is required<br />

to tend to zero for large values of t, since ˜ψ(t) has this property too. However,<br />

stability does not automatically result from the Padé approximation technique. Indeed,<br />

if emphasis is put on obtaining a good fit near t = 0 by choosing s 0 = ∞, it may easily<br />

happen that the resulting approximation becomes unstable: see again the 7th order approximation<br />

in Figure 4.3. The selection of a suitable point s 0 involves a choice between<br />

a good fit near t = 0 and stability, yielding a non-trivial problem which may be difficult<br />

to handle, depending on the wavelet at hand. In this respect it may be of interest to<br />

note that the 5th order approximation described in [54] and illustrated in Figure 4.3 was<br />

obtained with the choice s 0 = 0 which corresponds to a good fit in the time domain for<br />

large values of t as a result of the final value theorem (2.19). This however usually results<br />

in a poor fit at the beginning of the signal in the time domain, which is clearly visible in<br />

Figure 4.3 too.<br />

A third issue concerns the choice of the degrees m and n of the numerator and<br />

denominator polynomials of the rational approximation H(s). An unfortunate choice<br />

may yield an inconsistent system of equations or an unstable approximation. Changing<br />

m or n may solve this problem, but the converse may also happen: one may run into<br />

stability problems even if s 0 is left unchanged.<br />

In [19] an overview is given of various generalizations and extensions of Padé approximation<br />

which aim to deal with some of the problems just mentioned. For instance, it<br />

is possible to choose some or all of the poles of the rational approximation H(s) in advance.<br />

This offers a possibility to deal with the stability issue, but no clear theory exists<br />

on the optimal choices for these poles. Another generalization involves the possibility<br />

to use more than one interpolation point, which for instance offers a method to deal<br />

with the trade-off between s 0 = ∞ and s 0 = 0 in a more systematic way. Yet another<br />

possibility is to deal with many interpolation points s 0 , s 1 , . . . , s k and to require only a<br />

match between the values of H(s i ) and ˜Ψ(s i ) for i = 0, 1, . . . , k, no longer taking any<br />

derivatives at these points into account. One may even specify more interpolation points<br />

than the number of unknowns and use a linear least squares estimation technique to<br />

arrive at a unique solution. The advantage of such an approach is that one can optimize<br />

the function over a better distributed and controlled set of points than with classical<br />

Padé approximation. One may choose complex values here. A 9th order approximation<br />

of the Gaussian wavelet function obtained with this method is illustrated in Figure 4.3.<br />

This shows the feasibility of the approach, but for low order approximations the choice<br />

of interpolation points becomes more critical and the results are markedly less well.<br />

However, all of these techniques remain to have one important drawback in common:<br />

the quality of the approximation of the wavelet is not measured directly in the time<br />

domain but in the Laplace domain, and the criterion used does not allow for a direct<br />

interpretation in system theoretic terms.


4.4. L2 APPROXIMATION OF WAVELET FUNCTIONS 51<br />

4.4 L2 approximation of wavelet functions<br />

The theory of L 2 -approximation, can be formulated equally well in the time-domain and<br />

in the frequency domain, providing an alternative framework for studying the problem<br />

of wavelet approximation which offers a number of advantages over the Padé-approach.<br />

The advantages of this technique will be discussed, as well as an approach to find a<br />

suitable approximation.<br />

4.4.1 Wavelets and the L2 space<br />

On the conceptual level it is quite appropriate to use the L 2 -norm to measure the quality<br />

of an approximation h(t) of the function ˜ψ(t). Indeed, the very definition of the wavelet<br />

transform itself involves the L 2 -inner product between the signal f(t) and the mother<br />

wavelet ψ(t). It is also desirable that the approximation h(t) of ˜ψ(t) behaves equally well<br />

for all time instances t since h(t) is used as a convolution kernel with any arbitrary shift.<br />

This property holds naturally for L 2 -approximation, but it is not supported by the Padé<br />

approximation approach.<br />

Another advantage of L 2 -approximation is that it allows for a description in the time<br />

domain as well as in the Laplace domain, so that both frameworks can be exploited to<br />

develop further insight. According to Parseval’s identity [80] the squared L 2 -norm of the<br />

difference between ˜ψ(t) and h(t) can be expressed as:<br />

‖ ˜ψ − h‖ 2 =<br />

∫ ∞<br />

= 1<br />

2π<br />

−∞<br />

∫ ∞<br />

(<br />

˜ψ(t) − h(t)<br />

) 2<br />

dt, (4.8)<br />

−∞<br />

∣<br />

∣˜Ψ(iω) − H(iω) ∣ 2 dω. (4.9)<br />

Minimization of ‖ ˜ψ − h‖ is therefore equivalent to minimization of the L 2 -norm of the<br />

difference between the Laplace transforms ˜Ψ trunc (s) and H(s) over the imaginary axis<br />

s = iω. Note that this observation provides a rationale for the choice of interpolation<br />

points in a generalized Padé-approximation approach. In addition note that due to the<br />

causality of h(t) we can also write:<br />

‖ ˜ψ − h‖ 2 =<br />

∫ 0<br />

−∞<br />

= ζ ψ,t0 +<br />

( ) 2<br />

∫ ∞ ( ) 2<br />

˜ψ(t) dt + ˜ψ(t) − h(t) dt, (4.10)<br />

∫ ∞<br />

0<br />

0<br />

(<br />

˜ψ(t) − h(t)<br />

) 2<br />

dt, (4.11)<br />

so since ζ ψ,t0 does not depend on h(t) we may restrict the criterion to the positive real<br />

axis and the use of ˜ψ(t) in the optimization criterion is equivalent to the use of ˜ψ(t).<br />

One of the disadvantages of an L 2 -approximation approach is that there is a risk<br />

that the numerical optimization of ‖ ˜ψ − h‖ ends in a local, non-global optimum. Several<br />

global optimization techniques and software packages exist and if the problem can be<br />

rewritten in a specific form, then a global optimum can be found. However in general<br />

there is no guarantee whether a global optimum will be found or even whether it has


52 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

been found. Different starting points can give different local optima and thus can be<br />

used to find better solutions. Also, the outcomes of other approximation techniques can<br />

be used as starting points for L 2 -approximation.<br />

A question that arises is whether we can predict how good our approximation of the<br />

wavelet transformation is; i.e. not how well the basis functions are approximated, but<br />

how well the resulting detail coefficients match the detail coefficients that come from the<br />

actual transformation.<br />

Theorem 4.4.1. The squared error of the approximated wavelet transformation at any<br />

given point at any given scale |W (τ, σ) − Ŵ (τ, σ)| is no greater than the energy E s(t)<br />

of the signal s(t) multiplied by the squared error of the wavelet function approximation<br />

‖ ˜ψ(t) − h(t)‖.<br />

Proof. Denote the error of the wavelet approximation as:<br />

Given a scaled version of the wavelet:<br />

‖ ˜ψ(t) − h(t)‖ = ε (4.12)<br />

ψ τ,σ = 1 √ σ<br />

ψ<br />

( t − τ<br />

σ<br />

)<br />

, (4.13)<br />

and the resulting wavelet transform of signal s(t) at a given scale σ and place τ:<br />

W (τ, σ) = 〈s(t), ψ τ,σ (t)〉 =<br />

∫ ∞<br />

We can take their respective approximations:<br />

( )<br />

(<br />

1 t − τ 1<br />

√ ψ<br />

∼ √ h t 0 − t − τ )<br />

,<br />

σ σ<br />

σ σ<br />

( )<br />

1 −t + (τ + t0 σ)<br />

= √ h<br />

,<br />

σ σ<br />

−∞<br />

s(t)ψ τ,σ (t)dt. (4.14)<br />

= ˆψ τ,σ (t) (4.15)<br />

〈<br />

Ŵ (τ, σ) = s(t), ˆψ<br />

〉<br />

τ,σ (t) ,<br />

=<br />

∫ ∞<br />

−∞<br />

We can calculate the error of the approximation as:<br />

|W (τ, σ) − Ŵ (τ, σ)| = | 〈s(t), ψ τ,σ(t)〉 −<br />

s(t) ˆψ τ,σ (t)dt. (4.16)<br />

〈<br />

s(t), ˆψ<br />

〉<br />

τ,σ (t) |. (4.17)<br />

Due to the linearity of the inner product in each of the arguments:<br />

|W (τ, σ) −<br />

〈s(t), Ŵ (τ, σ)| = | ψ τ,σ (t) − ˆψ<br />

〉<br />

τ,σ (t) |. (4.18)<br />

Using the Cauchy-Schwarz inequality we obtain an upper bound:<br />

√<br />

|W (τ, σ) − Ŵ (τ, σ)| ≤ ‖s(t)‖ ‖ψ τ,σ(t) − ˆψ τ,σ (t)‖ = E s(t) ε. (4.19)


4.4. L2 APPROXIMATION OF WAVELET FUNCTIONS 53<br />

A number of remarks are in place here. First the upper bound depends on the energy<br />

of the total signal, whereas the wavelet functions have (effective) compact support. As<br />

a result the upper bound can be reduced by taking this into account. The upper bound<br />

then becomes dependent on the scale and properties of the wavelet. A second point is<br />

that the wavelets and their approximations can possess beneficial properties that further<br />

reduce the error, among which the possession of a vanishing moment as described in<br />

Section 4.4.3. This property avoids a bias in the approximated wavelet transform.<br />

4.4.2 Parameterization<br />

Particularly in the case of low-order approximation, the L 2 -approximation problem can<br />

be approached in a simple and straightforward way in the time domain. As is well known<br />

from linear systems theory (see, e.g., [62]) any strictly causal linear filter of finite order<br />

n can be represented in the time domain as a state-space system (A, B, C) as in (2.28).<br />

The impulse response function h(t) and its Laplace transform H(s) (i.e., the transfer<br />

function of the system) are then given by:<br />

h(t) = Ce At B, (4.20)<br />

H(s) = C(sI − A) −1 B. (4.21)<br />

For the generic situation of stable systems with distinct poles, the impulse response<br />

function h(t) is a linear combination of damped exponentials and exponentially damped<br />

harmonics. For low-order systems, this makes it possible to propose an explicitly parameterized<br />

class of impulse response functions among which to search for a good approximation<br />

of ˜ψ(t) as described in [67].<br />

The type of functions that come from (4.20) are the class of Bohl functions [97,<br />

Remark 3.5.3]. Bohl functions are sums of the products of polynomials and exponentials,<br />

which, in the real case, comes down to the sum of the product of polynomials, real<br />

exponentials, sines and cosines. The matrix exponential is defined as:<br />

e At def<br />

= I + At + A2 t 2<br />

+ A3 t 3<br />

+ . . . (4.22)<br />

2 3!<br />

The matrix A in (4.22) can be transformed, using a similarity transform, to Jordan form<br />

(a block-diagonal matrix).<br />

A = S −1 BS (4.23)<br />

e At = S −1 e Bt S, B in Jordan form (4.24)<br />

These blocks do not show interaction and each gives rise to a specific Bohl function.<br />

Scalar blocks obviously form pure exponentials. Products of polynomials and exponentials<br />

correspond to special Jordan structures in the matrix A and since they are not<br />

generic, they are not used, unless one has other reasons to consider them, in the parameterization.<br />

Typically, the functions that are used in the parameterization are damped


54 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

exponentials and exponentially damped sines and cosines. The latter can model the<br />

oscillatory behavior of wavelets.<br />

Sufficiently smooth continuous-time wavelets can be reasonably well approximated by<br />

strictly causal stable linear systems with distinct poles. The selected parameterization<br />

consists of damped exponentials and exponentially damped harmonics. The parameterized<br />

class of impulse response functions has the following form:<br />

h(t) = [ α 1 e p1t + . . . + α n e pnt] + [ β 1 e q1t sin(r 1 t) + γ 1 e q1t cos(r 1 t)+<br />

. . . + β m e qmt sin(r m t) + γ m e qmt cos(r m t) ] , (4.25)<br />

where the parameters p k and q k must be strictly negative for reasons of stability.<br />

For instance, if a 5 th order approximation is attempted, this parameterized class of<br />

functions h(t) may typically have the following form:<br />

h(t) = α 1 e p1t + β 1 e q1t sin(r 1 t) + γ 1 e q1t cos(r 1 t) + β 2 e q2t sin(r 2 t) + γ 2 e q2t cos(r 2 t), (4.26)<br />

Note that (4.26) can be rewritten as a sum of complex exponentials:<br />

h(t) = α 1 e p1t + η 1 e (q1+ir1)t + η ∗ 1e (q1−ir1)t + η 2 e (q2+ir2)t + η ∗ 2e (q2−ir2)t , (4.27)<br />

√ (<br />

with p k , q k ∈ R − β 2<br />

k<br />

, α k , β k , γ k , r k ∈ R, η k =<br />

+γ2 k<br />

2<br />

e i(arctan βk<br />

γ<br />

)+1 k R − (β k )π− π 2 ) and 1 A (x)<br />

is the indicator function: 1 A (x) = 1 iff x ∈ A.<br />

The parameterization has in this form some similarity to the classical problem of<br />

separation of exponentials [74, Chapter IV-23]. The problem deals with a function given<br />

in the following form:<br />

f(x) = ρ 1 e −λ1x + ρ 2 e −λ2x + . . . + ρ m e −λmx , (4.28)<br />

and aims to find the amplitudes ρ i and damping coefficients λ i . A difference with the<br />

problem at hand is that complex exponentials need to be found, and as a result the<br />

method would need modification.<br />

For the purpose of IC design it is useful to have a state-space representation (A, B, C)<br />

associated with (4.26) available. Such a representation is for instance provided by:<br />

⎛<br />

⎞ ⎛ ⎞<br />

p 1 0 0 0 0<br />

1<br />

0 q 1 r 1 0 0<br />

0<br />

A =<br />

0 −r<br />

⎜ 1 q 1 0 0<br />

, B =<br />

1<br />

,<br />

⎟ ⎜ ⎟<br />

⎝ 0 0 0 q 2 r 2 ⎠ ⎝0⎠<br />

0 0 0 −r 2 q 2 1<br />

C = ( α 1 β 1 γ 1 β 2 γ 2<br />

)<br />

. (4.29)<br />

Given the explicit form of the wavelet ˜ψ trunc (t) and the parameterized class of functions<br />

h(t), the L 2 -norm of the difference ˜ψ trunc (t) − h(t) can now be minimized in a<br />

straightforward way using standard numerical optimization techniques and software. In


4.4. L2 APPROXIMATION OF WAVELET FUNCTIONS 55<br />

[67, 66] a numerical approach was used that involves discretizing both the wavelet and<br />

the parameterized impulse response with a very fine mesh, and locally searching for the<br />

least-squares minimum of the difference between the two.The problem of avoiding local<br />

optima is discussed in Section 4.4.4. The negativity constraints on {p k } and {q k } which<br />

enforce stability are not difficult to handle for most optimization packages.<br />

4.4.3 Vanishing moments<br />

One common property of a wavelet function ˜ψ(t) that was undiscussed so far is that it<br />

must have at least one vanishing moment. Enforcing the first vanishing moment comes<br />

down to the requirement that the integral of the wavelet function is equal to zero:<br />

∫ ∞<br />

0<br />

˜ψ(t)dt = 0. (4.30)<br />

If this property is not shared by the approximation h(t), this will cause an unwanted<br />

bias in the approximation of the wavelet transform as can be seen in the simulation in<br />

Figure 4.4, where the h(t) obtained in this manner is denoted “L 2 normal” in the figure.<br />

This is likely to happen in a situation where a truncation error occurs.<br />

Proposition 4.4.2. The condition ∫ ∞<br />

0<br />

h(t)dt = 0 is equivalent to H(0) = 0.<br />

Proof.<br />

0 =<br />

∫ ∞<br />

0<br />

h(t)dt =<br />

∫ ∞<br />

0<br />

e −st h(t)dt, with s = 0 (4.31)<br />

= H(s) with s = 0 (4.32)<br />

= H(0). (4.33)<br />

In terms of linear filters, the property that the integral of the impulse response function<br />

h(t) is zero is equivalent to the property that the step response of the filter tends<br />

to zero for large t as follows from (2.27) and the final value theorem (2.19). Indeed, if<br />

the wavelet transform is computed for a step input signal, then a bias will be manifest if<br />

such a property is not satisfied.<br />

In terms of a state-space representation (A, B, C) we have that<br />

H(0) = −CA −1 B. (4.34)<br />

As an example, for the representation (4.29) it is not difficult to compute A −1 since it is<br />

block diagonal.<br />

⎛<br />

⎞<br />

p 1 0 0 0 0<br />

0 q 1 r 1 0 0<br />

A −1 =<br />

0 −r<br />

⎜ 1 q 1 0 0<br />

⎟<br />

⎝ 0 0 0 q 2 r 2 ⎠<br />

0 0 0 −r 2 q 2<br />

−1<br />

(4.35)


56 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

1<br />

0.8<br />

0.6<br />

0.4<br />

0.2<br />

0<br />

−0.2<br />

−0.4<br />

−0.6<br />

−0.8<br />

Simulation of various systems<br />

0.2*Dataset<br />

Regular wavelet<br />

L2 normal<br />

L2 step corr.<br />

−1<br />

0 50 100 150 200 250<br />

Figure 4.4: A scaled dataset is displayed along with its transformation with a wavelet.<br />

The black, thin, dashed line displays the output of a system that has the dataset as an<br />

input and attempts to approximate the wavelet transform without enforcing that the<br />

step response tends to zero for large t. The reader can observe that a bias manifests that<br />

correlates with the input data. The diamonds show the output of another system that<br />

does enforce the condition on the step response. For this system the bias is absent in the<br />

output data.<br />

⎛<br />

A −1 =<br />

⎜<br />

⎝<br />

1<br />

p 1<br />

0 0 0 0<br />

(<br />

)<br />

0 1 q 1 −r 1<br />

0 0<br />

0<br />

q1 2+r2 1 r 1 q 1 0 0<br />

(<br />

)<br />

0<br />

0 0<br />

1 q 2 −r 2<br />

0<br />

0 0<br />

q2 2+r2 2 r 2 q 2<br />

⎞<br />

⎟<br />

⎠<br />

(4.36)<br />

This yields the explicit condition:<br />

H(0) = α 1<br />

+ −β 1r 1 + γ 1 q 1<br />

p 1 q1 2 + + −β 2r 2 + γ 2 q 2<br />

r2 1 q2 2 + = 0 (4.37)<br />

r2 2<br />

If such an extra nonlinear condition is not conveniently handled by the optimization<br />

software, then it can easily be used to eliminate one of the variables from the problem:<br />

α 1 = −p 1<br />

(<br />

γ1 q 1 − β 1 r 1<br />

q 2 1 + r2 1<br />

+ γ )<br />

2q 2 − β 2 r 2<br />

q2 2 + r2 2<br />

(4.38)


4.4. L2 APPROXIMATION OF WAVELET FUNCTIONS 57<br />

Step and impulse response of L2 models with and without step correction<br />

1<br />

i.r. L2 normal<br />

i.r. L2 step corr<br />

s.r. L2 normal<br />

0.5<br />

s.r. L2 step corr.<br />

0<br />

−0.5<br />

−1<br />

0 5 10 15 20<br />

Figure 4.5: Impulse and step responses of the “normal” and “step corrected” L 2 -<br />

approximation<br />

4.4.4 Obtaining a good starting point<br />

When optimizing the parameterized class of impulse response functions to match a certain<br />

wavelet function in an L 2 -sense, it is a non-trivial task to find the global optimum. There<br />

generally exist a large number of local optima and it is very hard to find the global<br />

optimum. In fact the problem of global optimization in this setting is unsolved and no<br />

guarantees in general exist that a global optimum can or has been found. For specific<br />

problems these guarantees do exist and one can attempt to rewrite problems into such a<br />

form, however no suitable form has been found for the problem at hand. The optimization<br />

technique used in this study is to locally minimize a discretized least-squares criterion,<br />

since an l 2 -criterion is to be minimized. In order to prevent the optimization technique<br />

from terminating in an unsatisfactory local optimum, one can attempt to find a good<br />

starting point. The choice of the starting point can have a considerable impact on the<br />

solution found by the L 2 -approximation approach.<br />

A methodology is presented in [66] to obtain a good starting point for the L 2 -<br />

approximation approach in an automated fashion. One starts by constructing a highorder<br />

model and applying model reduction techniques such that an initial model with<br />

the appropriate order n req is obtained. A number of intermediate steps are required as<br />

illustrated in Figure 4.6 and listed below:<br />

1. Sampling the wavelet ˜ψ


58 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

High order<br />

discrete-time<br />

FIR-model<br />

Sampled<br />

wavelet<br />

function<br />

Intermediate<br />

order<br />

discrete-time<br />

IIR-model<br />

L 2<br />

approximation<br />

Initialization<br />

Intermediate<br />

order<br />

Continuoustime<br />

model<br />

2 3 4<br />

Balance and<br />

ZOH discrete to<br />

1<br />

Truncate<br />

Optimization<br />

Function<br />

6<br />

continuous time<br />

Balance and<br />

Truncate<br />

Low order<br />

continuoustime<br />

model<br />

5<br />

Restrictions<br />

Approximated<br />

wavelet<br />

function<br />

Model class<br />

Wavelet admissability<br />

conditions<br />

Figure 4.6: Automated approximation of functions<br />

The wavelet function ˜ψ is sampled with a sufficiently high resolution over a large<br />

enough time interval.<br />

2. Construction of a high order discrete-time FIR-model<br />

The sampled wavelet is used to construct a high order discrete-time FIR (or movingaverage)<br />

model, which has an impulse response that exactly matches the sampled<br />

wavelet.<br />

3. Conversion to an intermediate order discrete-time IIR-model<br />

The state-space model is balanced and truncated to yield an accurate reduced order<br />

discrete-time model, referred to as an intermediate order model.<br />

4. Conversion of the discrete-time IIR-model to a continuous-time IIRmodel<br />

The discrete-time model is then converted back to continuous-time, using the ZOHprinciple.<br />

Until here, all steps in the procedure have to be performed just once.<br />

5. Reduction of the continuous-time IIR-model to the desired lower order<br />

The intermediate order continuous-time model is reduced to a specified lower order<br />

n req , to be used as a starting point for the optimization technique in the next step.<br />

Various reduced orders can be attempted until a satisfactory result is obtained,<br />

steps 1-4 do not have to be repeated.<br />

6. L 2 -approximation of the wavelet function<br />

The low order model obtained in the previous step is used as a starting point for


4.4. L2 APPROXIMATION OF WAVELET FUNCTIONS 59<br />

solving the minimization problem described above under the constraint that for<br />

(4.34) it must hold that H(0) = 0, using an iterative local search optimization<br />

technique.<br />

Sampling the wavelet ˜ψ<br />

To obtain a good starting point one can start with a sampled version of the time-reversed<br />

and shifted wavelet ˜ψ(t) using sample intervals of size ∆t:<br />

f k = ˜ψ(k∆t), k = 0 . . . n. (4.39)<br />

The horizon n∆t has to be selected in such a way that stability is obtained. However<br />

the computational complexity grows with n. In the case of non-compactly supported<br />

wavelets the horizon can be selected by choosing a threshold α and setting n∆t close to<br />

the smallest t max for which | ˜ψ(t)| < α, ∀t > t max . One should select α in such a way that<br />

not only stability is obtained, but also the truncation error is sufficiently small. Stability<br />

is not guaranteed, but implicitly obtained in practice. For compactly supported wavelets<br />

n and ∆t should be selected in such a way that ˜ψ(n∆t) is well outside the support of<br />

the wavelet to ensure stability, but n is small enough to keep the calculations feasible. A<br />

typical value for ∆t is 0.01 and for n∆t a typical value is 20, depending on the decay of<br />

the function at hand.<br />

Construction of a high order discrete-time FIR-model<br />

The sequence {f k } will be used as the impulse response of a discrete-time system. Caution<br />

needs to be taken here since the impulse function in continuous-time is inherently different<br />

from the impulse response in discrete-time. The continuous-time impulse function is the<br />

Dirac delta-function (2.15), yielding an infinitesimally narrow, infinitely tall pulse which<br />

integrates to unity. On the other hand the discrete-time impulse can be associated<br />

(via zero-order hold) with a continuous time block function, called the Kronecker delta<br />

function (2.16). As a result a correction is required by filtering the sequence {f k } by a<br />

FIR block filter 1 2 + 1 2 z−1 to obtain { ˆf k }, which corresponds to a discrete-time impulse<br />

response. This sequence can be then be used to define a discrete-time system.<br />

The impulse response of the system under construction is chosen to be as follows:<br />

h[0] = 0, h[1] = ˆf 0 , h[2] = ˆf 1 , . . . , h[n + 1] = ˆf n . The impulse response of a discrete-time<br />

state-space system (A, B, C, d) is given by:<br />

h[k] =<br />

{ d k = 0<br />

CA k−1 B k > 0.<br />

(4.40)<br />

Therefore the required impulse response can be obtained by using a system M dh =<br />

(A dh , B dh , C dh , d dh ) in controllable companion form and a zero term d dh . The matrix


60 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

A dh then has the following form that shifts the state variables in time:<br />

⎛<br />

⎞<br />

0 . . . 0 0<br />

0<br />

A dh =<br />

⎜<br />

⎟<br />

⎝ I . ⎠<br />

0<br />

(4.41)<br />

The input is only fed to the first state and the vector C determines the output given<br />

the function that has to be fitted:<br />

⎛<br />

B dh = ⎜<br />

⎝<br />

1<br />

0<br />

.<br />

.<br />

0<br />

⎞<br />

⎟<br />

⎠<br />

( )<br />

C dh = ˆf0 ˆf1 . . . ˆfn<br />

d dh = 0<br />

(4.42)<br />

Eventually the model in (4.42) has to be converted to a low-order continuous time<br />

model. It may however not be easily converted directly to a continuous time model<br />

because A dh has all its poles at the origin, which makes it impossible to take a matrix<br />

logarithm as can be seen from (4.57). Therefore the model M dh is first reduced, then<br />

converted to continuous time and finally is reduced again to obtain a continuous-time<br />

model of the required low order.<br />

Conversion to an intermediate order discrete-time IIR-model<br />

To reduce the model M dh the balance and truncate procedure can be used for which the<br />

reader is referred to [89, 96, 121].<br />

M dh has to be balanced first. A stable, discrete-time system is balanced if the following<br />

system of equations hold where the first two equations are the well-known discretetime<br />

Lyapunov-Stein equations:<br />

P − AP A T = BB T (4.43)<br />

Q − A T QA = C T C (4.44)<br />

P = Q = diag{σ 1 , . . . , σ n } (4.45)<br />

for positive numbers σ k , known as the Hankel singular values of the system. The system<br />

does always admit a balanced realization and the Hankel singular values can be ordered<br />

in a decreasing fashion. Observe that (4.43) does only depend on A and B, and as a<br />

result not on the wavelet ˜ψ(t) that was fitted to M dh . It is not difficult to see that<br />

the matrix P , i.e., the controllability Grammian, corresponding to the model M dh is<br />

the identity matrix I. As a consequence P will be unaffected by orthogonal state-space<br />

transformations.


4.4. L2 APPROXIMATION OF WAVELET FUNCTIONS 61<br />

For (4.45) to hold, the matrix Q or the observability Grammian has to be diagonalized<br />

first by an orthogonal state-space transformation and then the system will be transformed<br />

by an additional diagonal state-space transformation to ensure P = Q.<br />

The diagonalization of Q is indeed possible with an orthogonal state-space transform<br />

of the form UQU T since the matrix Q is always positive definite in case of stability<br />

and minimality. The transform will have an influence on A, B and C , but not on P ;<br />

an identity matrix, as mentioned earlier. This makes it possible to transform Q with<br />

singular value decomposition.<br />

Ā dh = U T A dh U (4.46)<br />

¯B dh = U T B dh (4.47)<br />

¯C dh = C dh U (4.48)<br />

¯P = P = I (4.49)<br />

¯Q = U T QU (4.50)<br />

The orthogonal transformation matrix U will is such that ¯Q is a diagonal matrix. Condition<br />

(4.45) can now be ensured by an additional diagonal state-space transformation.<br />

The procedure discussed in [11] arrives at the same balanced realization in a slightly<br />

different way, by observing that for the controllable companion form realization of a<br />

moving-average system, the associated Hankel matrix H built from the finite impulse<br />

response 0, h[1], h[2], . . . , h[n] as<br />

⎛<br />

⎞<br />

h[1] h[2] h[3] . . . h[n]<br />

h[2] h[3] h[n] 0<br />

H =<br />

h[3] 0 .<br />

(4.51)<br />

⎜<br />

⎟<br />

⎝ . h[n] 0<br />

. ⎠<br />

h[n] 0 . . . . . . 0<br />

is diagonalized by the same matrix U. This avoids the computation of Q, of which the<br />

condition number is the square of the condition number of H.<br />

The controllability grammian P is defined as:<br />

∞∑<br />

P = A k BB T (A T ) k = CC T , (4.52)<br />

k=0<br />

where C = (B AB A 2 B . . .), i.e., the infinite controllability matrix. Similarly Q =<br />

OO T , where O T = (C T A T C T (A T ) 2 C T . . .), i.e. the observability matrix. Since<br />

h[k] = CA k−1 B and because A k−1 B = 0 for k > n, it is not necessary to consider an<br />

infinite matrix H. The Hankel matrix can be written in terms of the observability and<br />

controllability matrices as: HH T = OCC T O T = OO T , since P = I.<br />

After balancing ¯Q is a diagonal matrix with the singular values {σ 1 , . . . , σ n } of Q<br />

in decreasing order. Upon choosing a threshold ɛ σ , the state vector may be truncated,<br />

retaining as many n m state components as there are singular values above the threshold,<br />

yielding an n m th order discrete-time system M dm = (A dm , B dm , C dm , 0).


62 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

Conversion of the discrete-time IIR-model to a continuous-time IIR-model<br />

The matrix A dm will usually no longer contain zero eigenvalues and therefore the system<br />

M dm can be converted to continuous time as described in for example [115]:<br />

A dm = e Acm∆t (4.53)<br />

B dm =<br />

∫ ∆t<br />

0<br />

e Acmτ B cm dτ (4.54)<br />

C dm = C cm (4.55)<br />

with A dm , A cm , B dm and B cm further conveniently related by (see [118]):<br />

N =<br />

e N∆t =<br />

( )<br />

Acm B cm<br />

0 0<br />

( )<br />

Adm B dm<br />

0 1<br />

To compute the matrix N first diagonalize the matrix F =<br />

(4.56)<br />

(4.57)<br />

( )<br />

Adm B dm<br />

, using that<br />

0 1<br />

F ′ = V −1 F V = diag(k), (4.58)<br />

where V is the matrix with eigenvectors of F , k is the vector of eigenvalues of F and the<br />

function diag gives a diagonal matrix the given vector as its diagonal elements, produces<br />

a diagonal matrix F ′ if F . Next the matrix logarithm of F can be calculated as:<br />

N∆t = ln F = V ln F ′ V −1 (4.59)<br />

From (4.59) it can be seen that the logarithm of a matrix, of which the diagonal<br />

elements contain the diagonal entries of the matrix A m athrmdm has to be taken. Since<br />

the logarithm of zero is undefined one cannot transform matrices A dm that have zeros<br />

on the diagonal to continuous-time directly. Note that the controllable companion form<br />

with a matrix A m athrmdm as in (4.41) does have zeros on the diagonal and hence cannot<br />

be directly transformed to continuous-time.<br />

Reduction of the continuous-time IIR-model to the desired lower order<br />

The model M cm will not yet have the required order n req , therefore the continuous time<br />

model that has now been obtained has to be reduced even further with, for instance,<br />

the balance and truncate procedure. This procedure now involves the continuous-time<br />

Lyapunov equations, so that eventually a continuous-time model M cr of reduced order<br />

n req is obtained.


4.4. L2 APPROXIMATION OF WAVELET FUNCTIONS 63<br />

1<br />

0.5<br />

0<br />

Morlet wavelet approximation<br />

Morlet wavelet<br />

5 th order approximation<br />

7 th order approximation<br />

8 th order approximation<br />

-0.5<br />

-1<br />

0 2 4 6 8 10 12 14 16 18 20<br />

0.4<br />

0.2<br />

Morlet wavelet approximation error<br />

5 th order approximation error<br />

7 th order approximation error<br />

8 th order approximation error<br />

0<br />

-0.2<br />

-0.4<br />

0 2 4 6 8 10 12 14 16 18 20<br />

Figure 4.7: Automatic approximation of the Morlet wavelet<br />

L2-approximation of the wavelet function<br />

M cr is a single input - single output, strictly proper, asymptotically stable, continuoustime<br />

state-space system which will not in general be of the form as in (4.25), with a<br />

corresponding state-space realization of a form similarly as the example in (4.29) and<br />

will not obey the constraint as described in Section 4.4.3. To ensure that the step<br />

response of M cr tends to zero, the constant term in the numerator of the transfer function<br />

associated with M cr is simply set to zero to enforce a zero at zero. Next the system has<br />

to be brought in the model form as in (4.25). The complex eigenvalues associated with<br />

A cr will be ordered in complex pairs where the imaginary part is positive. The states<br />

corresponding to real eigenvalues follow last. The A matrix will have a block structure,<br />

from which the required parameters can be easily read off. Note that A cr fixes the number<br />

of complex pairs that the approximating model will have.<br />

In Figure 4.7 the approximation of the Morlet wavelet with systems of various reduced<br />

orders is illustrated. In the lower half of this figure the deviation from the ideal<br />

wavelet is shown. Since the “ideal” wavelet is truncated, it does not have a zero integral.<br />

The approximations however, do have this property, as described in Section 4.4.3, and<br />

therefore a perfect fit is not possible.


64 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

Abbreviation Matlab Release Function Algorithm<br />

lsqc2008bTrr R2008b “lsqcurvefit” “Algorithm”=“trust-region-reflective”<br />

fmin2008bTrr R2008b “fminsearch” “Algorithm”=“trust-region-reflective”<br />

lsqc14-3GN R14 SP3 “lsqcurvefit” “LargeScale”=“off”,“LevenbergMarquardt”=“off”<br />

lsqc2007bGN R2007b “lsqcurvefit” “LargeScale”=“off”,“LevenbergMarquardt”=“off”<br />

lsqc2008bGN R2008b “lsqcurvefit” “LargeScale”=“off”,“LevenbergMarquardt”=“off”<br />

fmin14-3NM R14 SP3 “lsqcurvefit” “LargeScale”=“off”,“LevenbergMarquardt”=“off”<br />

fmin2007bNM R2007b “fminsearch” “LargeScale”=“off”,“LevenbergMarquardt”=“off”<br />

fmin2008bNM R2008b “fminsearch” “LargeScale”=“off”,“LevenbergMarquardt”=“off”<br />

Table 4.2: Used Matlab settings and function for wavelet approximation. The settings<br />

“LargeScale”=“off”,“LevenbergMarquardt”=“off” attempts to use Gauss-Newton optimization<br />

for lsqcurvefit and Nelder-Mead simplex direct search for fminsearch. Under<br />

certain conditions the optimization may switch to Levenberg-Marquardt or to a largescale<br />

method. The lsqcurvefit function offers the possibility to specify upper and lower<br />

bounds for each of the parameters.<br />

Settings 12 th order 10 th order 8 th order 7 th order 6 th order 5 th order 3 rd order<br />

lsqc2008bTrr 0.0022 0.0009 0.0056 0.0070 0.0132 0.0475 0.3975<br />

fmin2008bTrr 0.0073 0.0068 0.0066 0.0073 0.0132 0.0475 0.3975<br />

lsqc14-3GN 0.0163 0.0198 0.0165 0.0173 0.0329 0.1168 0.9973<br />

lsqc2007bGN 0.0044 0.0103 0.0056 0.0070 0.0132 0.0475 0.3975<br />

lsqc2008bGN 0.0044 0.0103 0.0056 0.0070 0.0132 0.0475 0.3975<br />

fmin14-3NM 0.0164 0.0164 0.0173 0.0199 0.0329 0.1168 0.9973<br />

fmin2007bNM 0.0076 0.0068 0.0065 0.0077 0.0132 0.0475 0.3975<br />

fmin2008bNM 0.0073 0.0068 0.0066 0.0073 0.0132 0.0475 0.3975<br />

Table 4.3: Approximation of Gaussian wavelet with t 0 = 2.0 and Matlab settings as in<br />

Table 4.2.<br />

4.5 Empirical results<br />

The L 2 approximation approach for wavelets allows for the implementation of a wide<br />

variety of wavelets. In [52] it was discussed how complex wavelets can be implemented<br />

and in [65] a large number of wavelets are approximated with this approach. Besides<br />

the Gaussian wavelet, also the Morlet wavelet and the Mexican Hat wavelet, which<br />

are of interest for ECG processing [63, 3, 2, 20], have been approximated. Not only<br />

continuous wavelets have been approximated, but also discrete wavelets such as some of<br />

the Daubechies wavelets, provided that they are sufficiently smooth. The required order<br />

to obtain a satisfying approximation of these wavelets is quite high though.<br />

In [65] it is shown that a 4 th order L 2 approximation of a Gaussian wavelet outperforms<br />

a 5 th order Padé approximation, with respect to a multitude of evaluation methods,<br />

thus showing a significant improvement in performance. The approximated wavelets are<br />

shown in Figure 4.8.<br />

In order to illustrate how well various wavelets are approximated the l 2 norms of the<br />

approximation errors of various wavelets will be shown.<br />

As can be seen from Tables 4.3–4.8, accuracy does not always increase with the<br />

order. The use of a lower order solution as a starting point should give a performance<br />

increase in such cases. If a local optimum is nearby, the optimization may terminate<br />

in this local optimum. A way to overcome the latter problem is to introduce random<br />

disturbances on the starting point and use random starting point in the area around the<br />

deterministic starting points. Another observation is that with the same settings, the


4.5. EMPIRICAL RESULTS 65<br />

1<br />

(a)<br />

1<br />

(b)<br />

0.5<br />

0.5<br />

0<br />

0<br />

−0.5<br />

−0.5<br />

−1<br />

0 5 10 15 20<br />

−1<br />

0 5 10 15 20<br />

1<br />

(c)<br />

1<br />

(d)<br />

0.5<br />

0.5<br />

0<br />

0<br />

−0.5<br />

−0.5<br />

0 5 10 15 20<br />

−1<br />

0 5 10 15 20<br />

1<br />

(e)<br />

0.5<br />

(f)<br />

0.5<br />

0<br />

0<br />

−0.5<br />

−0.5<br />

−1<br />

0 5 10 15 20<br />

−1<br />

0 5 10 15 20<br />

Figure 4.8: Approximations of various wavelet functions. The thick gray line represents<br />

the wavelet function and the thin dashed black line represents the impulse response of<br />

the approximating system. The following wavelets were approximated with the specified<br />

orders n req and time shifts t 0 : (a) Gaussian n req = 5, t 0 = 2, (b) Morlet n req = 5, t 0 = 2.5,<br />

(c) Mexican Hat n req = 6, t 0 = 3.2, (d) Daubechies 3 n req = 6, t 0 = −1, (e) Daubechies<br />

7 n req = 8, t 0 = −4, (f) Coiflet 5 n req = 8, t 0 = −11.<br />

Settings 12 th order 10 th order 8 th order 7 th order 6 th order 5 th order 3 rd order<br />

lsqc2008bTrr 0.0026 1.5043 0.0277 0.0636 0.0926 0.2372 0.6318<br />

fmin2008bTrr 0.0030 0.0043 0.0277 0.0636 0.0926 0.2372 0.6398<br />

lsqc14-3GN 0.0026 0.0037 0.0277 0.0636 0.0926 0.2372 0.6318<br />

lsqc2007bGN 0.0026 0.0037 0.0277 0.0636 0.0926 0.2372 0.6318<br />

lsqc2008bGN 0.0026 0.0037 0.0277 0.0636 0.0926 0.2372 0.6318<br />

fmin14-3NM 0.0029 0.0043 0.0277 0.0636 0.0926 0.2372 0.6398<br />

fmin2007bNM 0.0029 0.0043 0.0277 0.0636 0.0926 0.2372 0.6318<br />

fmin2008bNM 0.0030 0.0043 0.0277 0.0636 0.0926 0.2372 0.6318<br />

Table 4.4: Approximation of Morlet wavelet with t 0 = 2.5 and Matlab settings as in<br />

Table 4.2.


66 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

Settings 12 th order 10 th order 8 th order 7 th order 6 th order 5 th order 3 rd order<br />

lsqc2008bTrr 0.0302 0.0051 0.0074 0.1447 0.1469 0.1782 0.9939<br />

fmin2008bTrr 0.1077 0.1320 0.1163 0.5844 0.1720 0.1721 0.5019<br />

lsqc14-3GN 0.0048 0.0082 0.0074 0.0189 0.0613 0.1782 0.5019<br />

lsqc2007bGN 0.3790 1.0549 1.1359 3.0077 0.1108 1.1009 0.5019<br />

lsqc2008bGN 0.6411 1.2933 1.0450 0.0189 0.0613 0.9974 0.5019<br />

fmin14-3NM 0.0060 0.0052 0.0083 0.0189 0.0613 0.1782 0.5019<br />

fmin2007bNM 0.1237 0.1871 0.2179 0.3885 0.4874 0.4850 0.5019<br />

fmin2008bNM 0.1077 0.1320 0.1163 0.5844 0.1720 0.1721 0.5019<br />

Table 4.5: Approximation of Mexican hat wavelet with t 0 = 3.2 and Matlab settings as<br />

in Table 4.2.<br />

Settings 12 th order 10 th order 8 th order 7 th order 6 th order 5 th order 3 rd order<br />

lsqc2008bTrr 0.8419 0.5685 0.9998 1.0001 1.0000 1.0001 1.0001<br />

lsqc14-3GN 0.1348 0.2785 0.6065 0.6065 0.6065 0.6086 1.0001<br />

lsqc2007bGN 0.8512 0.5835 0.8015 0.8477 1.0001 1.0001 1.0001<br />

lsqc2008bGN 0.8419 0.5685 0.9998 1.0001 1.0000 1.0001 1.0001<br />

fmin2007bNM ∗ 0.1440 1.0001 1.0000 0.9998 1.0001 1.0001 1.0001<br />

fmin2008bNM ∗ 0.5749 0.5728 0.9999 1.0000 1.0000 1.0000 1.0000<br />

Table 4.6: Approximation of Daubechies 3 wavelet with t 0 = 0.0 and Matlab settings<br />

as in Table 4.2. The ∗ indicates that a different starting point was used to obtain the<br />

approximation. For results close to one, the impulse response of the obtained systems<br />

looks very much like an impulse; a very undesirable approximation.<br />

Settings 12 th order 10 th order 8 th order 7 th order 6 th order 5 th order 3 rd order<br />

lsqc2008bTrr 0.0369 0.0429 0.0993 0.2430 0.2510 0.5241 0.8400<br />

fmin2008bTrr 0.0440 0.0449 0.0993 0.2430 0.2510 0.5241 0.8400<br />

lsqc14-3GN 0.0369 0.0429 0.0993 0.2430 0.2510 0.5241 0.8400<br />

lsqc2007bGN 0.0369 0.0429 0.0993 0.2430 0.2510 0.5241 0.8400<br />

lsqc2008bGN 0.0440 0.0449 0.0993 0.2430 0.2510 0.5241 0.8400<br />

fmin14-3GN 0.0440 0.0453 0.0993 0.2430 0.2510 0.5241 0.8400<br />

fmin2007bNM 0.0439 0.0443 0.0993 0.2430 0.2510 0.5241 0.8400<br />

fmin2008bNM 0.0440 0.0449 0.0993 0.2430 0.2510 0.5241 0.8400<br />

Table 4.7: Approximation of Daubechies 7 wavelet with t 0 = −4.0 and Matlab settings<br />

as in Table 4.2.<br />

Settings 12 th order 10 th order 8 th order 7 th order 6 th order 5 th order 3 rd order<br />

lsqc2008bTrr 0.0282 0.0706 0.1832 0.3510 0.4340 0.6335 0.8840<br />

fmin2008bTrr 0.0329 0.0730 0.1857 0.3510 0.0.4955 0.6335 0.8783<br />

lsqc14-3GN 0.0284 0.0706 0.1832 0.3510 0.4340 0.6335 0.8840<br />

lsqc2007bGN 0.0284 0.0706 0.1832 0.3510 0.4340 0.6335 0.8840<br />

lsqc2008bGN 0.0284 0.0706 0.1832 0.3510 0.4340 0.6335 0.8840<br />

fmin14-3GN 0.0329 0.0730 0.1857 0.3510 0.4955 0.6335 0.8783<br />

fmin2007bGN 0.0329 0.0730 0.1894 0.3510 0.4955 0.6335 0.8783<br />

fmin2008bGN 0.0329 0.0730 0.1857 0.3510 0.4955 0.6335 0.8783<br />

Table 4.8: Approximation of Coiflet 5 wavelet with t 0 = −11.0 and Matlab settings as<br />

in Table 4.2.


4.5. EMPIRICAL RESULTS 67<br />

0.8 time−reversed, time−shifted wavelet<br />

18th order continuous−time system<br />

7th order system ofter model reduction<br />

0.6<br />

7th order system ofter local search<br />

0.4<br />

0.2<br />

0<br />

−0.2<br />

−0.4<br />

−0.6<br />

0 5 10 15 20<br />

Figure 4.9: Approximations of Mexican Hat wavelet. The time-shifted, time-reversed<br />

Mexican Hat wavelet is displayed along its 18 th order approximation that results from<br />

step 4 of the procedure to find a suitable starting point. The 7 th order approximation from<br />

step 5 (which does not necessarily have an integral of 1) and the 7 th order approximation<br />

from step 6 are displayed as well and show a poor fit with respect to the Mexican Hat<br />

wavelet.<br />

acquired approximation using Matlab Release 14 Service Pack 3 is markedly better than<br />

with the newer Matlab versions.<br />

As can be seen from Table 4.5, the 7 th order approximation of the Mexican Hat<br />

wavelet, with the lsqcurvefit function and the Gauss-Newton algorithm, is dramatic.<br />

This approximation is illustrated in Figure 4.9. It is clear that the starting point for<br />

the L 2 wavelet design is very inconvenient. If one considers the Hankel singular values<br />

corresponding to the 18 th order approximation in Figure 4.10, one can observe that<br />

all but one states seem to contain relevant information about the system. This is<br />

further illustrated if the corresponding impulse responses are taken into consideration in<br />

Figure 4.11. The 17 th order system is still a good starting point and the quality quickly<br />

decreases. How dramatic the decrease is depends on a range of factors such as the wavelet<br />

at hand, the involved time shift, etc. And the effect of the quality of the starting point<br />

depends on the specific optimization surface at hand, the optimization routine and even<br />

the specific version of the software as becomes apparent from Tables 4.2–4.8.


68 CHAPTER 4. ANALOG IMPLEMENTATION OF WAVELETS<br />

1.2<br />

1<br />

Hankel singular value<br />

0.8<br />

0.6<br />

0.4<br />

0.2<br />

2 4 6 8 10 12 14 16 18<br />

state (ordered by Hankel singular value)<br />

Figure 4.10: Hankel singular values of the 18 th order approximation of the Mexican Hat<br />

wavelet that results from step 4 of the procedure to find a suitable starting point.


4.5. EMPIRICAL RESULTS 69<br />

0.8<br />

0.6<br />

0.4<br />

0.2<br />

0<br />

−0.2<br />

0.6<br />

0.4<br />

0.2<br />

0<br />

−0.2<br />

0.1<br />

0<br />

−0.1<br />

−0.2<br />

−0.3<br />

0<br />

−0.2<br />

−0.4<br />

order 18<br />

0 5 10 15 20<br />

order 15<br />

0 5 10 15 20<br />

order 12<br />

0 5 10 15 20<br />

order 9<br />

order 17<br />

0.8<br />

0.6<br />

0.4<br />

0.2<br />

0<br />

−0.2<br />

−0.4<br />

0 5 10 15 20<br />

0.6<br />

0.4<br />

0.2<br />

0<br />

−0.2<br />

0.4<br />

0.2<br />

0<br />

−0.2<br />

0<br />

−0.2<br />

−0.4<br />

order 14<br />

0 5 10 15 20<br />

order 11<br />

0 5 10 15 20<br />

order 8<br />

0.4<br />

0.2<br />

0<br />

−0.2<br />

order 16<br />

−0.4<br />

0 5 10 15 20<br />

0.2<br />

0.1<br />

0<br />

−0.1<br />

−0.2<br />

0.2<br />

0<br />

−0.2<br />

−0.4<br />

0<br />

−0.3347<br />

order 13<br />

0 5 10 15 20<br />

order 10<br />

0 5 10 15 20<br />

order 7<br />

−0.6<br />

0 5 10 15 20<br />

−0.6<br />

0 5 10 15 20<br />

−0.6695<br />

0 5 10 15 20<br />

Figure 4.11: Impulse responses of systems that are used as a starting point for the L 2<br />

approximation of the Mexican Hat wavelet.


Chapter 5<br />

Orthogonal wavelet design<br />

An important practical issue of interest is the choice of a wavelet basis [103]. When<br />

employing the Fourier transform the basis functions are fixed, i.e., sines and cosines.<br />

However for wavelets more freedom exists and the basis needs to satisfy the conditions<br />

associated with the wavelet framework. One possibility for wavelet selection is to use<br />

a parameterized wavelet that can be tuned for the application at hand [101]. Another<br />

possibility is to design a custom wavelet as is the case in this study [69]. Note that the<br />

current motivation for wavelet design is not to achieve a good quality lossy compression<br />

as is commonly seen, but feature detection, better representations and other signal<br />

processing applications. Earlier work on the design of wavelets matched the amplitude<br />

spectrum and the phase spectrum of a wavelet to a reference signal in the Fourier domain<br />

separately [24]. The authors of [46] describe an algorithm for the design of biorthogonal<br />

and semi-orthogonal wavelets. The design criterion comes down to the maximization of<br />

the energy in the approximation coefficients, however no clear motivation is provided by<br />

the authors. In [92] a parameterization of orthogonal wavelets, involving polyphase filters<br />

and the lattice structure as for example in [107] (see also Section 3.2), was discussed,<br />

however no clear design criterion was provided.<br />

In Section 5.1 two design criteria are introduced for discrete wavelets, that can be used<br />

to measure the quality of a discrete wavelet. This quality measure is with respect to a<br />

given signal and the key idea behind it is the maximization of the sparsity. In Section 5.2<br />

a parameterization of discrete-time orthogonal wavelets through polyphase filter banks<br />

with a lattice structure as in [92, 107] is discussed. Next the enforcement of additional<br />

vanishing moments and the use of the design criterion is discussed. Together with the<br />

results of Chapters 3 and 4 a complete procedure for orthogonal wavelet filter design<br />

is provided here. Using the wavelet equation (3.25) it is possible to compute a wavelet<br />

function associated with the designed wavelet filter bank. If this wavelet function is<br />

sufficiently smooth, it is possible to approximate the designed wavelet in analog circuits.<br />

The steps needed to come to an approximation for the implementation in analog circuits<br />

71


72 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

is illustrated in Figure 1.1.<br />

When multiple well-defined features in a signal need to be distinguished simultaneously,<br />

multiwavelets [76, 77, 107] can be employed. In the multiwavelet case multiple<br />

bases are used to span the l 2 space. In Section 3.4 it has been discussed how these multiwavelets<br />

can be parameterized in polyphase form and how the condition of orthogonality<br />

comes down to requiring that system in polyphase form is “lossless” as previously discussed<br />

in [116, 117, 107]. In order to design multiwavelets, a concrete parameterization of<br />

lossless systems is required, which is discussed in Section 5.3. For this parameterization<br />

in Section 5.3.1 the “tangential Schur algorithm” [55] is used. This parameterization is<br />

first provided for scalar wavelets, followed by the parameterization for multiwavelets. In<br />

this section the enforcement of a first vanishing moment, which is a requirement for a<br />

valid wavelet multiresolution structure, is discussed.<br />

5.1 Measures for the quality of a given representation<br />

When faced with an application, the choice of a wavelet is not a trivial task. A given<br />

wavelet can have desirable properties such as a large number of vanishing moments,<br />

optimal ratio in the Heisenberg uncertainty rectangle, linear phase, etcetera, but this<br />

may not give a direct answer to the question "what wavelet should be used?". In this<br />

work a quantitative measure is described to assess the quality of a given wavelet for a<br />

number of application types.<br />

For compression purposes a wavelet is good for a signal if the signal can be represented<br />

in the wavelet domain with only a few nonzero coefficients. This means that it has a sparse<br />

representation in the wavelet domain and is well located in both time and frequency. This<br />

is also a good property for detection purposes: if the wavelet is optimized for a certain<br />

feature, one then can identify it at a certain scale and relate it to a specific time. If one<br />

wants to detect anomalies, a deviation from this sparsity is a good indicator for this.<br />

The recursive application of the wavelet filter bank to the low-pass filter outputs<br />

(see Figure 3.3) gives a decomposition in terms of detail coefficients at all levels and in<br />

terms of approximation coefficient(s) at the coarsest level. In case orthonormal wavelets<br />

are employed, the energy of the signal is preserved in the detail and approximation<br />

coefficients (using the natation in Section 3.2.3) [107, page 27]:<br />

∑<br />

x 2 k =<br />

k<br />

j∑<br />

max<br />

j=1<br />

∑ (<br />

k<br />

b (j)<br />

k<br />

) 2 ∑ (<br />

+<br />

k<br />

a (jmax)<br />

k<br />

) 2<br />

(5.1)<br />

Note that this is the wavelet analogue of Parseval’s identity [80] as discussed on page 4.4.1.<br />

We introduce a vector w to consist of all these detail and approximation coefficients, i.e.<br />

w = (b (1)<br />

1 , . . . , b(1)<br />

k<br />

, b(2)<br />

1 , . . . , b(2) k<br />

, . . . , b(jmax) 1 , . . . , b (jmax)<br />

k<br />

, a (jmax)<br />

1 , . . . , a (jmax)<br />

k<br />

). (5.2)<br />

From (5.1) it follows that the l 2 -norm of this vector w (from now on one is implicitly<br />

referred to this vector when measuring the wavelet decomposition) is not an appropriate<br />

measure for the quality of the wavelet.


5.1. MEASURES FOR THE QUALITY OF A GIVEN REPRESENTATION 73<br />

The guiding principle to obtain sparsity that was proposed in [69] is to aim for the<br />

maximization of the variance [81]. This is worked out in the following two ways which<br />

are relevant for orthogonal wavelet transforms:<br />

1. maximization of the variance of the absolute values of the wavelet coefficients<br />

2. maximization of the variance of the squared wavelet coefficients<br />

The latter one maximizes the variance of the energy distribution over the detail and<br />

approximation coefficients at the various scales, and is shown below to correspond to<br />

maximization of the l 4 -norm. The former one corresponds to minimization of the l 1 -<br />

norm, which is a well-known criterion to achieve sparsity in various other contexts, see<br />

for example [33, 34, 22, 21]<br />

Theorem 5.1.1. Let w be the vector of the wavelet and the approximation coefficients<br />

as in (5.2), resulting from the processing of a signal x = (x 0 , x 1 , x 2 , . . . , x m ) by means of<br />

an orthogonal filter bank. Then:<br />

1. Maximization of the variance of the sequence of absolute values |w k | is equivalent<br />

to minimization of the l 1 -norm V 1 = ∑ m<br />

k=0 |w k|.<br />

2. Maximization of the variance of the sequence of energies |w k | 2 is equivalent to<br />

maximization of the l 4 -norm V 4 = ( ∑ m<br />

k=0 |w k| 4 ) 1/4 .<br />

Proof. We will prove 1. and 2. separately:<br />

1. As expressed in (5.1), the energy E in a signal is given by E = ∑ k |x k| 2 = ∑ k |w k| 2 ,<br />

irrespective of the choice of orthogonal wavelet basis. The variance of the vector of<br />

absolute values {|w k |} is given by<br />

∑k |w k| 2 (∑<br />

k<br />

−<br />

|w ) 2<br />

k|<br />

= E (∑<br />

m + 1 m + 1 m + 1 − k |w ) 2<br />

k|<br />

,<br />

m + 1<br />

in which E and m are constant. Hence maximization of this quantity is equivalent<br />

to minimization of V 1 .<br />

2. The variance of the vector of energies {|w k | 2 } is given by<br />

∑k |w k| 4 (∑k −<br />

|w k| 2 ) 2 ∑k<br />

= |w k| 4<br />

m + 1 m + 1 m + 1<br />

( ) 2 E<br />

−<br />

,<br />

m + 1<br />

in which E and m are constant. Hence maximization of this quantity is equivalent<br />

to maximization of the V 4 .


74 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

(a)<br />

(b)<br />

1350<br />

1300<br />

1250<br />

1200<br />

1150<br />

1100<br />

1050<br />

1000<br />

0.6<br />

0.4<br />

0.2<br />

0<br />

−0.2<br />

−0.4<br />

L 1<br />

wavelet func<br />

L 4<br />

wavelet func<br />

950<br />

50 100 150 200 250<br />

Time<br />

−0.6<br />

0 1 2 3 4 5<br />

Time<br />

D_6<br />

(c)<br />

700<br />

600<br />

D_6<br />

(d)<br />

400<br />

D_5<br />

500<br />

400<br />

D_5<br />

300<br />

200<br />

Wavelet scale<br />

D_4<br />

D_3<br />

D_2<br />

300<br />

200<br />

100<br />

0<br />

−100<br />

Wavelet scale<br />

D_4<br />

D_3<br />

D_2<br />

100<br />

0<br />

−100<br />

−200<br />

−300<br />

D_1<br />

50 100 150 200 250<br />

Time<br />

−200<br />

−300<br />

D_1<br />

50 100 150 200 250<br />

Time<br />

−400<br />

−500<br />

Figure 5.1: (a) Smoothed ECG beat of 256 samples. (b) Designed wavelets for ECG<br />

beat with filter length 8. (c) Wavelet decomposition of ECG beat with l 1 -optimized<br />

wavelet. In the top row the coarsest detail coefficients are displayed. The intensity of the<br />

blocks corresponds to the coefficient values. The lower rows show finer detail coefficients.<br />

The approximation coefficients have been omitted since their large values may hamper<br />

the readability of the detail coefficients. (d) Wavelet decomposition of ECG beat with<br />

l 4 -optimized wavelet.<br />

Example 5.1.2. As an example, take the smoothed ECG beat that is displayed in Figure<br />

5.1 along with the wavelets with filter length 8 that have been designed for this signal<br />

by optimizing the criteria V 1 and V 4 , respectively. In the lower part of the figure the<br />

wavelet decomposition of the signal using the respective designed wavelets is displayed.<br />

From Figure 5.1c and 5.1d it can be clearly seen that just a few wavelet coefficients are<br />

significantly larger than the rest as expected.<br />

5.2 Wavelet parameterization and design<br />

For wavelet design, it is inconvenient to take the set of filter coefficients directly as a<br />

search space for numerical optimization, on which additional constraints are imposed


5.2. WAVELET PARAMETERIZATION AND DESIGN 75<br />

(a)<br />

(b)<br />

a<br />

b<br />

c<br />

d<br />

+<br />

+ z -1<br />

Figure 5.2: Two examples of lattices<br />

to meet all of the necessary and desired conditions. Instead we aim to construct a<br />

parameterization which builds the desired properties in. In this research the choice<br />

is made to restrict to orthogonal wavelets. One of the advantages of this choice is<br />

that the inverse transform will be the adjoint of the forward transform H −1 = H † .<br />

Using the polyphase representation (Section 3.2.7) offers the advantage that the analysis<br />

and synthesis filters are adjacent, without down- and upsampling in between, making<br />

is easier to build in the required conditions for orthogonal wavelets from Section 3.2:<br />

normalization, double shift orthogonality and a first vanishing moment. The lattice<br />

structure [107] or all-pass systems are used to ensure that the filter is orthogonal as<br />

discussed below.<br />

5.2.1 Lattice structure<br />

If one has a orthonormal polyphase matrix H p (z), it can be expressed in lattice form.<br />

A lattice structure implements a filter H p (z) as a cascade of simple building blocks that<br />

each implement a single multiplication. In lattice form, wavelets can be implemented as a<br />

cascade of constant matrices and delays (see for example [107, section 4.5] and [116]). In<br />

Figure 5.2 two examples of lattices are displayed. The lattice in Figure 5.2a corresponds<br />

to the following polyphase filter matrix:<br />

[ ] a b<br />

. (5.3)<br />

c d<br />

And the lattice in Figure 5.2b with the matrix as in (5.6).<br />

As noted in for example [107] in order to have symmetry and consequently linear<br />

phase filters one can choose a = d and b = c. For orthonormal filters one chooses a = d<br />

and b = −c and, if desired, one appropriately normalizes the coefficients. The lattice<br />

structure is convenient since a cascade of linear phase filters is again linear phase and a<br />

cascade of orthogonal filters is again orthogonal. Depending on the design, each lattice<br />

will be constructed to be linear phase or orthogonal; quantization errors with respect to


76 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

2<br />

+<br />

cos 1<br />

z -1 -sin 1<br />

sin 1<br />

2<br />

+<br />

cos 1<br />

...<br />

...<br />

z -1<br />

+<br />

cos n<br />

-sin n<br />

sin n<br />

+<br />

cos n<br />

-1<br />

Figure 5.3: Lattice structure for a polyphase filter bank<br />

the parameters will not change this. Any 2-channel orthonormal filter can be expressed<br />

in lattice form [107].<br />

When choosing a = d and b = −c and using a suitable normalization, the lattice<br />

structure can also be implemented with a rotation matrix:<br />

[ ]<br />

cos θ − sin θ<br />

R =<br />

. (5.4)<br />

sin θ cos θ<br />

With this structure the 2n low pass filter coefficients ( as in (3.6)) c 0 , c 1 , . . . , c 2n−1<br />

can be reparameterized in terms of n new parameters θ 1 , θ 2 , . . . , θ n . Note that, unlike in<br />

[107], conventional rotation matrices are used.<br />

For k = 1, . . . , n, let<br />

[ ]<br />

cos θk − sin θ<br />

R(θ k ) =<br />

k<br />

, (5.5)<br />

sin θ k cos θ k<br />

and let<br />

Λ(z) =<br />

Then consider the 2 × 2 matrix product<br />

[ ] 1 0<br />

0 z −1 . (5.6)<br />

H(z) = R(θ n )Λ(z)R(θ n−1 )Λ(z) · · · R(θ 2 )Λ(z)R(θ 1 )Λ(−1). (5.7)<br />

The polyphase filter structure that follows from (5.7) is illustrated in Figure 5.3. With<br />

this lattice structure the orthogonality constraints from Section 3.2.2 are automatically<br />

satisfied [107].<br />

In Section 3 the convention is used that the energy of the wavelet low- and highpass<br />

filter coefficients is normalized to one:<br />

∑<br />

k c2 k = ∑ k d2 k<br />

= 1. A filter H(z) =<br />

( )<br />

∑ n−1 c2k c 2k+1<br />

k=0<br />

z −k that forms a wavelet filter bank will now be constructed. The<br />

d 2k d 2k+1<br />

coefficients c k and d k should therefore obey the conditions discussed in Section 3 (Normalization<br />

and double shift orthogonality and a vanishing moment). The coefficients<br />

d k will be related to the coefficients c k according to the alternating flip construction<br />

d k = (−1) k c 2n−1−k . For the sum of each phase (even and odd) of the filters the following<br />

holds [107]:


5.2. WAVELET PARAMETERIZATION AND DESIGN 77<br />

Proposition 5.2.1. For relation between 1) the sum of the even phase of the low-pass<br />

coefficients c 2(l−1) , 2) the sum of the odd phase of the low-pass coefficients c 2l−1 , 3)<br />

the sum of the even phase of the high-pass coefficients d 2(l−1) and 4) the sum of the<br />

odd phase of the high-pass coefficients d 2l−1 respectively, and the parameters θ k of the<br />

polyphase representation of wavelets in lattice form the following holds:<br />

(<br />

n∑<br />

n<br />

)<br />

∑<br />

c 2(l−1) = cos θ k (5.8)<br />

l=1<br />

k=1<br />

(<br />

n∑<br />

n<br />

)<br />

∑<br />

c 2l−1 = sin θ k<br />

l=1<br />

k=1<br />

(<br />

n∑<br />

n<br />

)<br />

∑<br />

d 2(l−1) = − sin θ k<br />

l=1<br />

k=1<br />

(5.9)<br />

(5.10)<br />

(<br />

n∑<br />

n<br />

)<br />

∑<br />

d 2l−1 = cos θ k . (5.11)<br />

l=1<br />

k=1<br />

Proof. We shall prove this proposition by induction. For n = 1 the wavelet filter coefficients<br />

become: c 0 = cos θ 1 , c 1 = sin θ 1 , d 0 = − sin θ 1 and d 1 = cos θ 1 for which the<br />

theorem clearly holds. Now assume that for n = p the proposition holds, making the<br />

∑ p<br />

following equations true:<br />

l=1 c 2(l−1) = cos ( ∑ p<br />

k=1 θ k), ∑ p<br />

l=1 c 2l−1 = sin ( ∑ p<br />

∑ k=1 θ k),<br />

p<br />

l=1 d 2(l−1) = − sin ( ∑ p<br />

k=1 θ k) and ∑ p<br />

l=1 d 2l−1 = cos ( ∑ p<br />

k=1 θ k). Now consider Figure<br />

5.3. From the lattice structure it can be seen that for n = p + 1 it holds for the even<br />

phase that:<br />

(<br />

n∑<br />

p∑<br />

)<br />

( p∑<br />

)<br />

c 2(l−1) = cos θ p+1 cos θ k − sin θ p+1 sin θ k<br />

k=1 k=1<br />

l=1<br />

= cos<br />

( ∑p+1<br />

)<br />

θ k .<br />

k=1<br />

For the odd phase it holds that:<br />

(<br />

n∑<br />

p∑<br />

)<br />

( p∑<br />

)<br />

c 2l−1 = sin θ p+1 cos θ k + cos θ p+1 sin θ k<br />

k=1 k=1<br />

l=1<br />

= sin<br />

( ∑p+1<br />

)<br />

θ k .<br />

k=1<br />

For the even phase of the high-pass filter it holds that<br />

(<br />

n∑<br />

p∑<br />

)<br />

( p∑<br />

)<br />

d 2(l−1) = − sin θ p+1 cos θ k − cos θ p+1 sin θ k<br />

k=1 k=1<br />

l=1<br />

= − sin<br />

( ∑p+1<br />

)<br />

θ k .<br />

k=1


78 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

For the odd phase it holds that:<br />

(<br />

n∑<br />

p∑<br />

)<br />

( p∑<br />

)<br />

d 2l−1 = − sin θ p+1 sin θ k + cos θ p+1 cos θ k<br />

k=1 k=1<br />

l=1<br />

= cos<br />

( ∑p+1<br />

)<br />

θ k .<br />

k=1<br />

which completes the proof.<br />

Due to (5.8) and (5.9) the identity<br />

( ∑<br />

l<br />

c 2(l−1)<br />

) 2<br />

+<br />

c 2l−1<br />

) 2<br />

= 1 (5.12)<br />

( ∑<br />

l<br />

holds.<br />

From Figure 5.3, the relation between the parameters θ k and the filter coefficients<br />

can be observed. If the filter coefficients for a lattice structure with n = p are known,<br />

then their relation with the filter coefficients for the case n = p + 1 and the additionally<br />

involved parameter θ p+1 can be easily found.<br />

Observation 5.2.2. Suppose that the p × 2 array with filter coefficients for the case<br />

n = p is known to be<br />

⎛<br />

⎞<br />

c 0 c 1<br />

c 2 c 3<br />

⎜<br />

⎟<br />

⎝ . . ⎠ .<br />

c 2p−1<br />

c 2(p−1)<br />

Then the (p + 1) × 2 array with filter coefficients for the case n = p + 1 is:<br />

⎛<br />

⎞<br />

c 0 cos θ p+1 c 0 sin θ p+1<br />

c 2 cos θ p+1 − c 1 sin θ p+1 c 2 sin θ p+1 + c 1 cos θ p+1<br />

c 4 cos θ p+1 − c 3 sin θ p+1 c 4 sin θ p+1 + c 3 cos θ p+1<br />

.<br />

.<br />

⎜<br />

⎟<br />

⎝c 2(p−1) cos θ p+1 − c 2p−3 sin θ p+1 c 2(p−1) sin θ p+1 + c 2p−3 cos θ p+1 ⎠<br />

−c 2p−1 sin θ p+1 c 2p−1 cos θ p+1<br />

as follows from the lattice structure in Figure 5.2.<br />

⎛<br />

=<br />

⎜<br />

⎝<br />

⎞<br />

c 0 0<br />

c 2 c 1<br />

c 4 c 3<br />

R(θ p+1 ) T (5.13)<br />

. .<br />

⎟<br />

c 2p−3 ⎠<br />

0 c 2p−1<br />

c 2(p−1)


5.2. WAVELET PARAMETERIZATION AND DESIGN 79<br />

Example 5.2.3. As an example the low-pass coefficients for the filters of orders n = 2 and<br />

n = 3 are given:<br />

n = 2 c 0 = cos θ 1 cos θ 2<br />

c 1 = cos θ 1 sin θ 2<br />

c 2 = − sin θ 1 sin θ 2<br />

c 3 = sin θ 1 cos θ 2<br />

n = 3 c 0 = cos θ 1 cos θ 2 cos θ 3<br />

c 1 = cos θ 1 cos θ 2 sin θ 3<br />

c 2 = − sin θ 1 sin θ 2 cos θ 3 − cos θ 1 sin θ 2 sin θ 3<br />

c 3 = − sin θ 1 sin θ 2 sin θ 3 + cos θ 1 sin θ 2 cos θ 3<br />

c 4 = − sin θ 1 cos θ 2 sin θ 3<br />

c 5 = sin θ 1 cos θ 2 cos θ 3<br />

In order to avoid a bias in the wavelet transform and in order to make it possible<br />

to obtain a well-defined scaling and wavelet function (see Section 3.2.4 and [107, section<br />

6.1]), the wavelet filter must possess at least one vanishing moment. As explained in<br />

Section 3.2.5, when a wavelet has vanishing moments, this offers a number of advantages<br />

such as giving the wavelet regularity and smoothness. In terms of the filter coefficients<br />

this comes down to the constraint:<br />

∑<br />

c 2(l−1) − ∑ c 2l−1 = 0. (5.14)<br />

l<br />

l<br />

In terms of the parameters θ 1 , . . . , θ n this comes down to the following:<br />

Theorem 5.2.4. Consider a polyphase filter in lattice structure with the parameters<br />

θ 1 , . . . , θ n . For this filter to have at least a single vanishing moment the condition is<br />

[107, Theorem 4.6]:<br />

n∑<br />

θ k = π + l2π, l ∈ Z (5.15)<br />

4<br />

k=1<br />

Proof. In terms of the filter coefficients the condition comes down to ∑ n<br />

l=1 c 2(l−1) =<br />

∑ n<br />

l=1 c 2l−1. From Proposition 5.2.1 it follows that this is equivalent with the condition:<br />

cos ( ∑ n<br />

k=1 θ k) = sin ( ∑ n<br />

k=1 θ k). This condition is obviously satisfied if ∑ n<br />

k=1 θ k = π 4 +<br />

lπ, l ∈ Z. However, when using the sign convention that ∑ 2n−1<br />

k=1 c k = √ 2, it is required<br />

that cos ( ∑ n<br />

k=1 θ k) = sin ( ∑ n<br />

k=1 θ k) = 2√ 1 2, which excludes half of the previous solutions.<br />

∑ n<br />

Therefore the constraint becomes:<br />

k=1 θ k = π 4 + l2π, l ∈ Z, which completes the<br />

proof.<br />

5.2.2 Enforcing additional vanishing moments<br />

To impose additional vanishing moments is possible, but substantially more complicated<br />

since the parameters θ 1 , . . . , θ n enter the equations in a nonlinear way. As an example<br />

the condition to have a second vanishing moment is now worked out.


80 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

Theorem 5.2.5. In order for the wavelet low-pass filter resulting from the lattice structure<br />

to have two vanishing moments, for the parameters θ 1 , . . . , θ n it must hold in addition<br />

to (5.15) that<br />

( ( ))<br />

n−1<br />

∑<br />

k∑<br />

cos 2 θ l + 1 2 = 0 (5.16)<br />

k=1<br />

l=1<br />

Proof. The condition of the first vanishing moment H 1 (1) = 0 with the corresponding<br />

condition on the parameters {θ k } in (5.15) is assume to hold. For the second vanishing<br />

( ) ( )<br />

1<br />

moment the additional condition H 1(1) ′ = 0 must hold. Since H(z 2 H0 (z)<br />

)<br />

z −1 = ,<br />

H 1 (z)<br />

we have that: ( )<br />

( ) ( )<br />

d H0 (z)<br />

1<br />

0<br />

= 2zH ′ (z 2 )<br />

dz H 1 (z)<br />

z −1 + H(z 2 )<br />

−z −2 . (5.17)<br />

From (5.7) we find:<br />

( 1<br />

√ )<br />

H(1) =<br />

2 2<br />

1<br />

2√<br />

2<br />

√ √<br />

1<br />

2 2 −<br />

1<br />

, (5.18)<br />

2 2<br />

so that the last term of (5.17) at z = 1 yields:<br />

( ( )<br />

0 −<br />

1<br />

H(1) = 2√<br />

2<br />

√ . (5.19)<br />

−1)<br />

2<br />

Now we determine H ′ (z):<br />

H ′ (z) = R(θ n )Λ ′ (z)R(θ n−1 )Λ(z)R(θ n−2 )Λ(z) . . . Λ(z)R(θ 1 )Λ(−1)<br />

+ R(θ n )Λ(z)R(θ n−1 )Λ ′ (z)R(θ n−2 )Λ(z) . . . Λ(z)R(θ 1 )Λ(−1)<br />

+ R(θ n )Λ(z)R(θ n−1 )Λ(z)R(θ n−2 )Λ ′ (z) . . . Λ(z)R(θ 1 )Λ(−1)<br />

+ . . .<br />

1<br />

2<br />

+ R(θ n )Λ(z)R(θ n−1 )Λ(z)R(θ n−2 )Λ(z) . . . Λ ′ (z)R(θ 1 )Λ(−1) (5.20)<br />

Since we are interested in H ′ (z 2 ) at z = 1 we note that Λ(1) = I, Λ(z) becomes identity<br />

( ) 0 0<br />

and Λ ′ (1) = . The following is obtained:<br />

0 −1<br />

n−1<br />

∑<br />

(( (∑ n cos<br />

H ′ (1) =<br />

l=k+1 θ l)<br />

sin (∑ n<br />

k=1<br />

l=k+1 θ )<br />

l<br />

⎛<br />

)<br />

⎛<br />

n−1<br />

∑<br />

= ⎝<br />

k=1<br />

(<br />

⎝ cos ∑k<br />

l=1 θ l<br />

( ∑k<br />

)<br />

sin<br />

l=1 θ l<br />

sin<br />

(∑ n<br />

l=k+1 θ l)<br />

sin<br />

( ∑k<br />

l=1 θ l<br />

− cos (∑ n<br />

l=k+1 θ l)<br />

sin<br />

( ∑k<br />

l=1 θ l<br />

− sin (∑ n<br />

l=k+1 θ ) ) ( )<br />

l 0 0<br />

cos (∑ n<br />

l=k+1 θ )<br />

l 0 −1<br />

( ∑k<br />

) ⎞<br />

− sin<br />

l=1 θ ( ) ⎞<br />

l<br />

( ∑k<br />

) ⎠ 1 0<br />

⎠<br />

cos<br />

l=1 θ 0 −1<br />

l<br />

)<br />

− sin (∑ (<br />

n<br />

l=k+1 θ ∑k<br />

)<br />

l)<br />

cos<br />

l=1 θ l<br />

)<br />

cos (∑ (<br />

n<br />

l=k+1 θ ∑k<br />

)<br />

l)<br />

cos<br />

l=1 θ l<br />

⎞<br />

⎠ .<br />

(5.21)


5.2. WAVELET PARAMETERIZATION AND DESIGN 81<br />

( 1<br />

Now H ′ (1) is determined:<br />

1)<br />

( 1<br />

H ′ (1)<br />

1)<br />

=<br />

=<br />

=<br />

=<br />

=<br />

⎛<br />

n−1<br />

∑<br />

⎝ sin (∑ n<br />

l=k+1 θ l) ( ( ∑k<br />

) (<br />

sin<br />

l=1 θ ∑k<br />

)) ⎞<br />

l − cos<br />

l=1 θ l<br />

k=1<br />

cos (∑ n<br />

l=k+1 θ l) ( ( ∑k<br />

) (<br />

cos<br />

l=1 θ ∑k<br />

)) ⎠<br />

l − sin<br />

l=1 θ ,<br />

l<br />

⎛√ (∑<br />

n−1<br />

∑<br />

n 2 sin<br />

l=k+1<br />

⎝<br />

θ ) ( ∑k<br />

) ⎞<br />

l sin<br />

l=1 θ l − π 4<br />

√ (∑ (<br />

n<br />

k=1 2 cos<br />

l=k+1 θ l)<br />

sin<br />

π<br />

4 − ∑ ) ⎠<br />

k<br />

l=1 θ ,<br />

l<br />

⎛<br />

n−1<br />

∑ − √ (<br />

2 sin 2 π<br />

⎝<br />

4 − ∑ ) ⎞<br />

k<br />

l=1 θ l<br />

√ (<br />

k=1 2 cos<br />

π<br />

4 − ∑ ) (<br />

k<br />

l=1 θ π<br />

l sin<br />

4 − ∑ ) ⎠<br />

k<br />

l=1 θ ,<br />

l<br />

⎛ (<br />

n−1<br />

∑<br />

⎝ −√ 2 sin 2 π<br />

4 − ∑ ) ⎞<br />

k<br />

l=1 θ l<br />

√ (<br />

1<br />

k=1 2 2 sin<br />

π<br />

2 − 2 ∑ ) ⎠<br />

k<br />

l=1 θ ,<br />

l<br />

⎛ (<br />

n−1<br />

∑<br />

⎝ −√ 2 sin 2 π<br />

4 − ∑ ) ⎞<br />

k<br />

l=1 θ l<br />

√<br />

2 cos<br />

(2 ∑ ) ⎠<br />

k<br />

l=1 θ . (5.22)<br />

l<br />

k=1<br />

1<br />

2<br />

Due to Theorem 5.2.4 it also holds that:<br />

( 1<br />

H ′ (1) =<br />

1)<br />

n−1<br />

∑<br />

k=1<br />

( 1<br />

√ ( ∑ n<br />

2 2 cos 2<br />

l=k+1 θ l<br />

( ∑<br />

1<br />

n<br />

2√<br />

2 sin 2<br />

l=k+1 θ l<br />

) ) −<br />

1<br />

2√<br />

2<br />

) . (5.23)<br />

Since the condition for the second vanishing moment is D ′ (1) = 0 the bottom row of<br />

(5.17) needs to be equal to zero. From (5.17), (5.19) and (5.22) we find that the condition<br />

for the second vanishing moment comes down to:<br />

which completes the proof.<br />

( (<br />

n−1<br />

∑<br />

cos 2<br />

k=1<br />

))<br />

k∑<br />

θ l + 1 = 0, (5.24)<br />

2<br />

l=1<br />

The reader should note that this condition cannot be conveniently enforced by eliminating<br />

a parameter. As a concrete example, consider the case with three parameters<br />

(n = 3) of which θ 1 is fixed in order to enforce a first vanishing moment. In this case the<br />

condition (5.16) becomes:<br />

cos (2θ 2 + 2θ 3 ) + cos (2θ 3 ) + 1 = 0. (5.25)<br />

2<br />

Given that the condition for the first vanishing moment is satisfied it may happen that<br />

it is not possible to satisfy the condition for the second vanishing moment by fixing<br />

another parameter. For n = 3, the curve implicitly parameterized by (5.24) is displayed<br />

in Figure 5.4.


82 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

Figure 5.4: Curve where (5.25) is satisfied. The dot marks the location of the<br />

Daubechies 3 wavelet in the parameter space.<br />

5.2.3 Design and optimization<br />

The parameterized lattice structure discussed previously allows for the design of wavelets,<br />

where orthogonality is built into the model class. The parameters θ 1 , . . . , θ n can be chosen<br />

to optimize a certain design goal as discussed in Section 5.1. The issues involved with<br />

optimizing over the previously discussed parameter space are discussed in this section.<br />

Local optima<br />

With the criteria described in Section 5.1, combined with the theory of polyphase filters<br />

and the lattice structure, it is possible to construct an orthogonal wavelet that is optimal<br />

in the sense of a chosen criterion function. To compute this orthogonal wavelet,<br />

a local search method can be employed to search the parameter space θ 1 , . . . , θ n for an<br />

optimal wavelet. As often happens with local search techniques, there is a risk that the<br />

optimization may terminate in a local optimum, as will be illustrated in the following<br />

example.<br />

Example 5.2.6. The smoothed ECG signal from Figure 5.1a is used as a prototype signal.<br />

The number of parameters is chosen to be n = 3, yielding two degrees of freedom and a<br />

resulting filter of length 6. As design criterion the l 1 -norm of the wavelet coefficients of<br />

the wavelet transform is minimized. A single vanishing moment was enforced. During<br />

the experimentation we discovered a number of local minima, as illustrated in Figures<br />

5.5, 5.6 and 5.7.


5.2. WAVELET PARAMETERIZATION AND DESIGN 83<br />

−1.5<br />

−1<br />

−0.5<br />

0<br />

0.5<br />

1<br />

1.5<br />

−1.5<br />

−1<br />

x 10 4 θ 3<br />

2<br />

Criterion value<br />

1.5<br />

1<br />

θ 2<br />

−0.5<br />

0<br />

0.5<br />

1<br />

1.5<br />

Figure 5.5: The existence of local minima<br />

In Figure 5.6 and Figure 5.7 the local optima are further investigated in detail. In the<br />

upper left corner of the figure the location of the local optima is displayed in a contour<br />

plot of the l 1 -criterion over the free parameter space. In the other subfigures the wavelet<br />

and scaling function corresponding to the optima in the contour plot are displayed. From<br />

this figure it can be seen that some of these local optima can be quite similar, where one<br />

is a flipped or shifted version of another one, yielding almost the same result. Note that<br />

local optimum 3 is so close to number 8 that their difference is negligible. It is contained<br />

in the same gully due to the periodicity of the parameter space. The fact that this<br />

local optimum is found from both sides of the parameter space substantiates that this<br />

is indeed a local optimum and not a border artefact. If one considers the low-pass filter<br />

[−0.0016, −0.1225z −1 , 0.2217z −2 , 0.8359z −3 , 0.4870z −4 , −0.0062z −5 ] and the high-pass<br />

filter [−0.0062, −0.4870z −1 , 0.8359z −2 , −0.2217z −3 , −0.1225z −4 , 0.0016z −5 ], associated<br />

with this local optimum, it is easy to see that these effectively have a filter of length four,<br />

since leading or trailing coefficients are not significantly large.<br />

Observation 5.2.7. An interesting consequence of the preceding example is that if one<br />

compares the significant filter coefficients of local optima 3 and 8 with the low-pass filter<br />

of the Daubechies 2 wavelet: [−0.1294, 0.2241z −1 , 0.8365z −2 , 0.4830z −3 ] and high-pass


84 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

2<br />

1<br />

1<br />

Ψ<br />

φ<br />

2<br />

1<br />

Ψ<br />

φ<br />

2<br />

0<br />

0<br />

−1<br />

0 1 2 3 4 5<br />

2<br />

1<br />

0<br />

Ψ<br />

φ<br />

−1<br />

0 1 2 3 4 5<br />

2<br />

0<br />

φ<br />

−2<br />

0 1 2 3 4 5<br />

2<br />

1<br />

0<br />

Ψ<br />

φ<br />

−1<br />

0 1 2 3 4 5<br />

3<br />

5<br />

7<br />

Ψ<br />

−1<br />

0 1 2 3 4 5<br />

2<br />

1<br />

0<br />

−1<br />

0 1 2 3 4 5<br />

2<br />

1<br />

0<br />

−1<br />

0 1 2 3 4 5<br />

2<br />

1<br />

0<br />

Ψ<br />

φ<br />

−1<br />

0 1 2 3 4 5<br />

4<br />

6<br />

8<br />

Ψ<br />

φ<br />

Ψ<br />

φ<br />

Figure 5.6: Local optima in detail. The wavelet and scaling functions corresponding to<br />

the local optima found in the design of a wavelet for the signal in Figure 5.1a using three<br />

free parameters.


5.2. WAVELET PARAMETERIZATION AND DESIGN 85<br />

1.5<br />

1<br />

4<br />

6<br />

8<br />

7<br />

0.5<br />

θ 3<br />

0<br />

−0.5<br />

−1<br />

1<br />

2<br />

−1.5<br />

5<br />

3<br />

−1.5 −1 −0.5 0 0.5 1 1.5<br />

θ 2<br />

Figure 5.7: The location of the local optima with respect to the free parameters θ 2 and<br />

θ 3 as in Figure 5.6 is displayed.


86 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

Start<br />

Generate<br />

random<br />

θ 2 ,…,θ n<br />

Set θ 1 such<br />

that<br />

∑θk= π/4+ k2π<br />

Determine H 0 (z)<br />

and H 1 (z) from<br />

θ 1 ,…,θ n<br />

Determine<br />

improved<br />

θ 2 ,…,θ n<br />

Decompose<br />

the signal x<br />

with filters<br />

H 0 (z) and H 1 (z)<br />

Maximum<br />

iterations<br />

reached?<br />

Solution<br />

converged?<br />

Determine criterion<br />

value V 1 or V 4<br />

Figure 5.8: Algorithm to find an optimal wavelet<br />

filter of the Daubechies 2 wavelet: [−0.4830, 0.8365z −1 , −0.2241z −2 , −0.1294z −3 ], they<br />

are nearly identical. This observation provides a rationale for the use of the Daubechies<br />

2 wavelet for ECG processing.<br />

It can be concluded that local optima are present in the search space associated with<br />

the parameterization described in this section and the criteria described in Section 5.1.<br />

The problem of finding a global optimum in the presence of local optima is very hard.<br />

There are a number of ways in which these local optima can be avoided. There are a<br />

number of techniques that attempt to find a global optimum in the presence of local<br />

optima such as for example simulated annealing [71], genetic algorithms [42] and branch<br />

and bound methods [6]. However in general these methods cannot guarantee that a<br />

global optimum can be found and even that an optimum that has been found is indeed<br />

a global or local optimum. In light of these drawbacks and the applications at hand it<br />

was decided to run the local search algorithm multiple times with a number of different,<br />

randomly selected starting points as a heuristic of finding a global optimum.<br />

Optimization algorithm<br />

As discussed in the previous section a local search technique was employed and this search<br />

algorithm was used a number of times with random starting points. This approach is<br />

illustrated in Figure 5.8, where the blocks with solid borders visualize the local search<br />

method and the blocks with the dashed borders represent the loop with random starting<br />

points. This approach is implemented in Matlab, using the implementation of the Nelder-<br />

Mead direct search (simplex) algorithm [73] that is available in the function fminsearch


5.2. WAVELET PARAMETERIZATION AND DESIGN 87<br />

of the Matlab Optimization Toolbox.<br />

5.2.4 Experimentation<br />

In order to verify the effectiveness of the proposed approach a number of tests were<br />

conducted.<br />

Test for reproducibility<br />

In order to validate the approach, first consider that if a signal has a sparse representation<br />

in the wavelet domain (when transformed with a certain wavelet), this wavelet is<br />

likely to give a good performance for the criteria discussed earlier with respect to the<br />

wavelet decomposition of the signal in question. Note that although both the criteria V 1<br />

and V 4 aim at maximizing the variance, they still may give different results. However<br />

when starting with a sparse representation in the wavelet domain and reconstructing an<br />

artificial signal, the chance that differently shaped wavelets perform equally well becomes<br />

smaller. As a validation of the approach, we start with a sparse set of wavelet coefficients<br />

and wavelet filter. The associated time series is used as an input for the wavelet design<br />

procedure, and it is verified whether the wavelet filter that is discovered as and optimum<br />

is equal to the wavelet filter used to reconstruct the signal from the wavelet coefficients.<br />

As example of the test, first random parameters θ 2 and θ 3 are selected and θ 1 is<br />

fixed such that (5.15) holds and the corresponding sequences of scaling and wavelet filter<br />

coefficients (resp. H 0 and H 1 ) are determined. Next the full wavelet decomposition of a<br />

signal of length 256 is considered and only a few of these wavelet coefficients are assigned<br />

non-zero values. Carrying out the synthesis algorithm, we are constructing a signal in the<br />

wavelet domain. These wavelet coefficients are considered to be a sparse representation<br />

of an unknown signal x in terms of H 0 and H 1 . Now the signal x is determined by taking<br />

the sparse wavelet decomposition and performing an inverse wavelet transform on it with<br />

orthogonal wavelet filters H 0 and H 1 , i.e. the signal x is reconstructed from the sparse<br />

wavelet representation using the wavelet filters H 0 and H 1 .<br />

We now have a signal x for which we know that H 0 and H 1 will give a sparse representation,<br />

thus this wavelet filter bank should perform well with respect to the optimization<br />

criterion and there should be a reasonably large chance that this wavelet filter bank is<br />

found as an optimum. To verify this, the signal x was taken and the optimal wavelet was<br />

determined with the approach described in this report. Optionally, before determining<br />

the optimal wavelet, additive white Gaussian noise was added to the signal x in order to<br />

test the effect of noise on the approach. The testing procedure is visualized in Figure 5.9<br />

and the results for 250 trials are displayed in Table 5.1. If the l 1 -norm of the difference<br />

of the six actual and calculated filter coefficients is larger than 0.01, then the test is<br />

considered to be a failure.<br />

Without noise the original wavelet was recovered in virtually all cases as can be seen<br />

in Table 5.1. If the SNR decreases a clear sparse representation of the signal in the<br />

wavelet domain may no longer exist but, in particular for the l 4 criterion, the original


88 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

Select random<br />

parametersθ 1<br />

, θ 2<br />

, θ 3<br />

Determine wavelet<br />

filter from parameters<br />

Compare<br />

wavelet filters<br />

Select small number of<br />

nonzero wavelet coefficients<br />

Reconstruct signal<br />

in time domain with<br />

inverse wavelet<br />

transform<br />

Find optimal<br />

wavlet with<br />

n=3 for given<br />

signal<br />

1<br />

1<br />

Optionally: Add white noise<br />

Figure 5.9: Test for reproducibility<br />

SNR l 1 recovery rate l 4 recovery rate<br />

∞ 99.2% 100%<br />

40 dB 99.6% 99.6%<br />

20 dB 32.4% 44.4%<br />

10 dB 3.60% 9.60%<br />

0 dB 0.00% 1.60%<br />

Table 5.1: Percentage of times that a wavelet that provided a sparse representation of<br />

the prototype signal was recovered. Noise was added to the prototype signal yielding a<br />

Signal to Noise Ratio as indicated in the table.<br />

wavelet was still found as an optimum in many cases. Taking the presence of local optima<br />

in consideration the results are fair.<br />

Filter length versus criterion value<br />

The filter length of the optimized wavelet is twice the number of parameters n. A large n<br />

gives more freedom to optimize the wavelet with respect to the criterion since in order to<br />

enforce the required first vanishing moment, a single parameter has to be fixed. However,<br />

as n increases, the complexity of the optimization increases as well, and it becomes more<br />

difficult to avoid local optima. The “curse of dimensionality” applies here: choosing<br />

random initial points to cover the domain becomes more and more cumbersome; there<br />

is an exponential increase of effort. To determine the effect of the choice of n on the<br />

criterion value that is achieved during the optimization a test was conducted. For the<br />

data in Figure 5.1 the optimal wavelet was determined for n = 1, 2, . . . , 25. For each<br />

choice of the number of parameters, i.e. for each choice of the filter length, a thousand<br />

random starting points were generated and the best criterion value in terms of both the


5.3. MULTIWAVELET PARAMETERIZATION AND DESIGN 89<br />

(a) − l 1<br />

criterion<br />

(b) − l 4<br />

criterion<br />

10500<br />

−1800<br />

10000<br />

−1850<br />

9500<br />

−1900<br />

fval<br />

9000<br />

fval<br />

−1950<br />

8500<br />

8000<br />

7500<br />

−2000<br />

−2050<br />

5 10 15 20 25<br />

Number of parameters<br />

5 10 15 20 25<br />

Number of parameters<br />

Figure 5.10: criterion value vs n for the l 1 criterion in (a) and for the l 4 criterion in<br />

(b). Note that the values of the l 4 criterion have been multiplied with −1 such that the<br />

optimization problem becomes a minimization problem.<br />

l 1 and l 4 criterion were stored. The results are illustrated in Figure 5.10. From this figure<br />

it can be concluded that it is not beneficial for the given signal and criterion to use more<br />

than eight parameters. The increase in criterion value for a large number parameters is<br />

a consequence of the existence of local optima.<br />

5.3 Multiwavelet parameterization and design<br />

If one wants to discriminate between two features in a signal with a different morphology<br />

then multiwavelets (Section 3.4) [119, 116, 76, 77, 107] are a powerful tool. As discussed<br />

in Section 3.4 multiwavelets are a number of mutually orthogonal wavelet functions. In<br />

that section it was discussed that for orthogonal multiwavelets with compact support in<br />

a polyphase representation the polyphase matrices become lossless [116, 117] and FIR.<br />

These orthogonal multiwavelets can be designed using a parameterization as lossless systems,<br />

by jointly optimizing each multiwavelet function with respect to different segments<br />

of a prototype signal as first discussed in [95]. The class of orthogonal multiwavelets with<br />

compact support is parameterized with the results from [55]. The same design criteria<br />

as discussed in Section 5.1 are used for multiwavelet design.<br />

First, in the following subsection, the parameterization for (multi)wavelets using lossless<br />

systems will be discussed. This employs interpolation theory and the theory of balanced<br />

realizations of lossless systems by means of the tangential Schur algorithm [55].<br />

The approach consists of a number of steps. A key observation to start from is that<br />

a balanced realization of a lossless system has a unitary realization matrix. One way<br />

of parameterizing unitary realization matrices is by building them as a product of elementary<br />

unitary matrices. For this purpose, mappings F U,V are introduced. These


90 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

mappings are a special simple kind of so called “linear fractional transformations”, which<br />

are extensively used in interpolation theory. The tangential Schur algorithm allows one<br />

to parameterize lossless transfer functions through a recursive procedure which involves<br />

linear fractional transformations; this procedure was cast in state-space terms in [55]. To<br />

parameterize real FIR lossless polyphase matrices, a number of special choices for the<br />

parameters in the tangential Schur algorithm can be made, which substantially simplify<br />

the expressions.<br />

Then, in the next subsection, the use of the tangential Schur algorithm [55] to parameterize<br />

scalar wavelets is explained. It is shown how to recover the product formula<br />

(5.7) encountered in the lattice filter implementation by making suitable choices in the<br />

algorithm. The current set-up allows us to generalize this to the case of orthogonal multiwavelets<br />

with compact support. We also manage to build in a first “balanced vanishing<br />

moment”, which is much needed for practical purposes. This balanced vanishing moment<br />

was previously enforced in the literature as a constraint, but in this work a parameterization<br />

for multiwavelets with a built-in balanced vanishing moment is presented. In the<br />

last subsection a design approach for multiwavelets is discussed in which the multiwavelet<br />

parameterization is used to arrive at concrete results.<br />

5.3.1 Parameterization of lossless systems<br />

As discussed in Section 3.4 the condition of orthogonality for polyphase multiwavelet<br />

filters comes down to requiring that H p (z) is lossless, i.e., stable all-pass. We recall that<br />

for an arbitrary all-pass system G(z) of size p × p it holds that:<br />

G(e iω ) † G(e iω ) = I p , ∀ω ∈ R. (5.26)<br />

Conversely, if (5.26) holds then G(z) is all-pass. If G(z) satisfies the following properties:<br />

G(z) † G(z) = I p for |z| = 1, (5.27)<br />

G(z) † G(z) ≥ I p for |z| < 1, (5.28)<br />

G(z) † G(z) ≤ I p for |z| > 1, (5.29)<br />

where each two equations imply the remaining third one, then G(z) is also stable, thus<br />

lossless.<br />

For any proper rational matrix function R(z) we introduce the following notation:<br />

R † (z) = R(z) † = R † 0 + R† 1 z−1 + R † 2 z−2 + . . . (5.30)<br />

For z = e iω it holds that R † (z −1 ) = R(z) † . Consequently an all-pass system G(z) has<br />

the property that an inverse exists and is given by:<br />

G(z) −1 = G † (z −1 ), (5.31)<br />

for z = e iω , hence for all complex z by analytical continuation (see for example [74]).


5.3. MULTIWAVELET PARAMETERIZATION AND DESIGN 91<br />

When considering a 2r × 2r lossless polyphase multiwavelet filter H p (z) with multiplicity<br />

r, as described in Section 3.4, it can be partitioned as:<br />

(<br />

)<br />

H 0,e (z) H 0,o (z)<br />

H p (z) =<br />

, (5.32)<br />

H 1,e (z) H 1,o (z)<br />

with H 0,e (z), H 0,o (z), H 1,e (z) and H 1,o (z) the even and odd parts of the low- and highpass<br />

multiwavelet filters, respectively. Due to (3.58) and (5.31) we have that:<br />

(<br />

) (<br />

) ( )<br />

H 0,e (z) H 0,o (z) H † 0,e (z−1 ) H † 1,e (z−1 ) I r 0<br />

H 1,e (z) H 1,o (z) H † 0,o (z−1 ) H † =<br />

. (5.33)<br />

1,o (z−1 ) 0 I r<br />

These conditions must be exactly met in order to generate a valid orthogonal multiwavelet<br />

structure. Although they can be enforced by adding them as constraints to an<br />

optimization routine during the multiwavelet design, this is not convenient as they are<br />

nonlinear and they will often be satisfied merely approximately. A better and more elegant<br />

way is to find a parameterization in which these conditions are built-in. Procedures<br />

for the recursive construction of lossless systems exist in the literature; the procedures<br />

for building all-pass systems from [55] will be used in the remainder of this chapter to<br />

parameterize orthogonal multiwavelets.<br />

A construction method for p × p lossless systems of order n will now be described.<br />

When compact support is required, corresponding to the FIR property of the filters<br />

H 0 (z), H 1 (z) as well as H p (z), special choices can be made to parameterize this subclass<br />

too. Additional properties such as a (balanced) vanishing moment can also be incorporated.<br />

However, the problem of building in vanishing moments of higher order, and in<br />

fact of finding the appropriate balancing conditions, is currently not completely solved<br />

[95, 27, 77, 76, 109, 23, 108].<br />

A key result which underlies the parameterization procedure for lossless systems is the<br />

following. From [55, Proposition 3.2] it follows that if the realization matrix associated<br />

with a realization (A, B, C, D) of some rational function G(z) is unitary, then G(z) will<br />

be lossless. If A is additionally asymptotically stable then (A, B, C, D) will actually be<br />

a minimal and balanced representation. Recall that a balanced realization satisfies the<br />

Lyapunov-Stein equations (4.43)-(4.45), while in the balanced lossless case it holds that<br />

P = Q = I n .<br />

Theorem 5.3.1. Let (A, B, C, D) (with dimensions n × n, n × p, p × n and p × p<br />

respectively) be a balanced state-space realization of an n th order lossless p-input, p-output<br />

transfer function G(z) = D + C (zI n − A) −1 B. Then the (p + n) × (p + n) realization<br />

( ) D C<br />

matrix R =<br />

is unitary.<br />

B A<br />

Conversely, let R be such a unitary block-partitioned matrix, then G(z) = D +<br />

C (zI n − A) −1 B is p × p lossless of degree less than or equal to n. It is of degree n<br />

if and only if A is asymptotically stable, in which case the realization is balanced.<br />

For a proof see [56, 55]. This theorem allows the problem of parameterizing lossless<br />

functions to be studied in terms of unitary realization matrices, associated with balanced


92 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

realizations of lossless functions. Such approach was first carried out in [56] for scalar<br />

(1 × 1) lossless transfer functions, where a balanced canonical form was constructed with<br />

a corresponding upper triangular reachability matrix. The associated realization matrix<br />

allows for a factorization into a product of n more simple unitary building blocks.<br />

This approach was generalized in [55] to multivariable (p × p) lossless functions transfer<br />

functions of degree n. There, it is described how to parameterize such lossless systems<br />

using the tangential Schur algorithm. The associated canonical forms (A, B, C, D) are<br />

parameterized with the associated Schur parameter vectors, while other quantities serve<br />

to index local coordinate charts. In the multivariable case instead of unitary matrix<br />

multiplications, so-called linear fractional transformations are employed to set up a recursive<br />

procedure to construct the realization matrix. It was also found for which choices<br />

in the tangential Schur algorithm this procedure in fact reduces to unitary matrix multiplications<br />

for realization matrices, which is useful from a practical viewpoint. These<br />

procedures will now be described, where the following definitions are helpful.<br />

On the level of transfer functions, for a proper rational matrix G(z), we introduce a<br />

mapping F U,V :<br />

F U,V : G(z) → ˜G(z) (5.34)<br />

defined by:<br />

where<br />

(<br />

F 1 (z)<br />

F 3 (z)<br />

˜G(z) = F 1 (z) + F 2(z)F 3 (z)<br />

z − F 4 (z) , (5.35)<br />

⎛<br />

)<br />

F 2 (z)<br />

= V<br />

F 4 (z) ⎜<br />

⎝<br />

1 0 . . . 0<br />

0<br />

.<br />

0<br />

G(z)<br />

⎞<br />

⎟<br />

⎠ U † , (5.36)<br />

and U and V are (p + 1) × (p + 1) matrices, F 1 (z) is a p × p matrix, F 2 (z) a p × 1 vector,<br />

F 3 (z) a 1 × p vector and F 4 (z) a scalar.<br />

Additionally, it is proven in [55] that if G(z) is proper then ˜G(z) will be well-defined.<br />

Note that lossless transfer functions are always proper.<br />

When using unitary matrices U and V , the mapping F U,V (G(z)) takes lossless functions<br />

to lossless functions. In fact, it constructs a lossless transfer function ˜G(z) with an<br />

order at most one larger than the order of the lossless system G(z).<br />

Theorem 5.3.2. Let U, V be two unitary (p + 1) × (p + 1) matrices and let G(z) be<br />

p × p lossless of degree n. Then ˜G(z) = F U,V (G(z)) is also lossless of degree less than or<br />

equal to n + 1.<br />

For a proof see [55].<br />

On the level of balanced state-space realizations of lossless transfer functions, the<br />

mapping F U,V can be implemented as described by the following result (see again [55]):<br />

Theorem 5.3.3. Let U, V be two unitary (p+1)×(p+1) matrices and let G(z) be p×p<br />

lossless of degree n, with a balanced realization (A, B, C, D). Then ˜G(z) = F U,V (G(z))


5.3. MULTIWAVELET PARAMETERIZATION AND DESIGN 93<br />

has the realization (Ã, ˜B, ˜C, ˜D) given by:<br />

( ) ⎛ ⎞<br />

1 0 0 (<br />

V 0 ⎜ ⎟<br />

⎝ 0 D C ⎠<br />

0 I n<br />

0 B A<br />

) (<br />

U † 0 ˜D ˜C<br />

=<br />

0 I n<br />

˜B Ã<br />

)<br />

, (5.37)<br />

where Ã, ˜B, ˜C, ˜D respectively have dimensions (n + 1) × (n + 1), (n + 1) × p, p × (n + 1)<br />

and p × p; hence the state-space dimension is n + 1. It is minimal and balanced if and<br />

only if à happens to be asymptotically stable.<br />

This theory allows for a parameterization of lossless systems in state-space by iteratively<br />

applying unitary matrix multiplications. It is also possible to parameterize lossless<br />

systems in terms of transfer functions by using so-called “linear fractional transformations”<br />

as in the more general framework of interpolation theory, more specifically using<br />

the tangential Schur algorithm. A lossless function g(z) increases in McMillan degree by<br />

exactly one into a lossless function ˜G(z), in each step of this tangential Schur algorithm.<br />

The tangential Schur algorithm employs linear fractional transformations (LFT) associated<br />

with J-inner matrices. In [55] LFTs are considered of the form:<br />

(<br />

Θ1 Θ<br />

where Θ =<br />

2<br />

T Θ : G → (Θ 4 G + Θ 3 )(Θ 2 G + Θ 1 ) −1 , (5.38)<br />

Θ 3 Θ 4<br />

)<br />

is a 2p×2p block-partitioned rational matrix function of McMillan<br />

degree m, with blocks of size p × p and G is a p × p rational matrix of McMillan degree<br />

n.<br />

Theorem 5.3.4. If G(z) is p × p lossless of degree n and Θ(z) is 2p × 2p J-inner of<br />

degree m, then ˜G(z) = T Θ(z) (G(z)) is p × p lossless of degree less than or equal to n + m.<br />

Here the function Θ(z) is called J-inner if it holds that:<br />

Θ(z) † JΘ(z) = J for |z| = 1, (5.39)<br />

Θ(z) † JΘ(z) ≤ J for |z| < 1, (5.40)<br />

Θ(z) † JΘ(z) ≥ J for |z| > 1, (5.41)<br />

( )<br />

Ip 0<br />

with J =<br />

. For a proof see [55, Proposition 4.4]. Note that the inverse of a<br />

0 −I p<br />

J-inner function Θ(z) is given by:<br />

Θ(z) −1 = JΘ † (z −1 )J. (5.42)<br />

In the tangential Schur algorithm, a rational p×p lossless function of McMillan degree<br />

n is reduced in n recursion steps to such a function of degree 0; that is, to a constant<br />

unitary matrix. The degree is reduced by m = 1 in each step and the J-inner functions<br />

associated with these steps are also of McMillan degree one. Such function are called<br />

“elementary J-inner factors”.


94 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

The elementary J-inner factors that are employed in [55] are those which have their<br />

pole outside the closed unit disk at z = ¯w −1 . They are represented as:<br />

⎛<br />

⎛(<br />

Θ(u, v, w, ξ, M)(z) =<br />

⎜<br />

⎝ I 2p + ⎝<br />

z−w<br />

1− ¯wz<br />

(<br />

ξ−w<br />

1− ¯wξ<br />

[ ] [ ⎞<br />

) ⎞ u u<br />

J<br />

v v]†<br />

) − 1⎠<br />

(‖u‖ 2 − ‖v‖ 2 ) ⎟ M. (5.43)<br />

⎠<br />

In this expression, the matrix M ∈ 2p×2p is a J-unitary constant matrix, u ∈ C p×1 with<br />

‖u‖ = 1 is a normalized direction vector, v ∈ C p×1 with ‖v‖ < 1 is a Schur vector, w ∈ C<br />

with |w| < 1 is an interpolation point and ξ ∈ C with |ξ| = 1 is a normalization point on<br />

the unit circle. Note that Θ(u, v, w, ξ, M)(ξ) = M.<br />

From [55, Proposition 5.3] we have that:<br />

Theorem 5.3.5. Let G(z) be p × p lossless of McMillan degree n, then<br />

˜G(z) = T Θ(u,v,w,ξ,H)(z) (G(z)) is p × p lossless of McMillan degree n + 1, satisfying the<br />

interpolation condition ˜G( ¯w −1 )u = v.<br />

Conversely from [55, Proposition 5.4] we have:<br />

Theorem 5.3.6. Let ˜G(z) be p × p lossless of McMillan degree n + 1, let w ∈ C be such<br />

that |w| < 1, let u ∈ C p+1 with ‖u‖ = 1 be such that v := ˜G( ¯w −1 )u satisfies ‖v‖ < 1.<br />

Let ξ ∈ C satisfy |ξ| = 1 and M ∈ C 2p×2p be constant J-unitary. Then there exists a<br />

unique lossless p × p function G(z) of McMillan degree n such that ˜G(z) can be obtained<br />

as ˜G(z) = T Θ(u,v,w,ξ,H)(z) (G(z)).<br />

In each step of the parameterization procedure, the order is increased by one, by<br />

choosing values for w, u, ξ and M and by letting v vary over the space of vectors in C p<br />

with ‖v‖ < 1. The interpolation conditions make clear that this gives us coordinate charts<br />

for the space, i.e. the manifold, of p × p lossless functions of degree n. Multiple charts<br />

are needed to cover the manifold; it can be shown that no single Euclidean coordinate<br />

chart is capable of covering the manifold entirely.<br />

In our implementation, a lossless system is preferably built in state-space using a<br />

recursion, i.e., the space of p × p lossless function is parameterized by application of<br />

the reversed tangential Schur algorithm, carried over to the state-space level, starting<br />

from a constant p × p unitary matrix. To achieve this, we describe which choices in the<br />

tangential Schur algorithm allow an LFT ˜G(z) = T Θ(u,v,w,ξ,H)(z) (G(z)) to be represented<br />

as a mapping ˜G(z) = F u,v (G(z)) for suitable unitary matrices U and V . This is discussed<br />

in [55, Theorem 6.4].<br />

Theorem 5.3.7. If T Θ(u,v,w,ξ,M)(z) coincides with a mapping F U,V with U and V unitary,<br />

then<br />

( ) P 0<br />

Θ(u, v, w, ξ, M)(z) = H(uv † )S u,w (z)H(wuv † )<br />

, (5.44)<br />

0 Q


5.3. MULTIWAVELET PARAMETERIZATION AND DESIGN 95<br />

for some u, v ∈ C p with ‖u‖ = 1, ‖v‖ < 1, some w ∈ C with |w| < 1, and some p × p<br />

unitary matrices P and Q. Equivalently it holds that<br />

( ) P 0<br />

M = H(uv † )S u,w (ξ)H(wuv † )<br />

. (5.45)<br />

0 Q<br />

( )<br />

( )<br />

1 0<br />

In that case one can take U = Û<br />

and V =<br />

0 P<br />

̂V 1 0<br />

, where<br />

0 Q<br />

Û =<br />

̂V =<br />

⎛<br />

⎜<br />

⎝<br />

⎛<br />

⎜<br />

⎝<br />

√<br />

1−|w| 2<br />

√<br />

1−|w| 2 ‖v‖ 2 u I p − (1 + w√ 1−‖v‖ 2<br />

√<br />

1−|w| 2 ‖v‖ 2 )uu†<br />

w √ 1−‖v‖ 2<br />

√<br />

1−|w|2 ‖v‖ 2 √<br />

1−|w| 2<br />

√<br />

1−|w|2 ‖v‖ 2 u†<br />

√<br />

1−|w| 2<br />

√<br />

1−|w|2 ‖v‖<br />

√ v 2 1−‖v‖ 2<br />

√<br />

1−|w| 2 ‖v‖ 2<br />

√<br />

1−‖v‖ 2<br />

I p − (1 − √ ) vv†<br />

1−|w|2 ‖v‖ 2<br />

−<br />

√<br />

1−|w| 2<br />

√<br />

1−|w| 2 ‖v‖ v† 2<br />

‖v‖ 2<br />

⎞<br />

⎟<br />

⎠ , (5.46)<br />

⎞<br />

⎟<br />

⎠ . (5.47)<br />

In this theorem, the notation H(E) is used to denote the Halmos extension of a<br />

strictly contractive p × p matrix E:<br />

( )<br />

(Ip − EE † ) −1/2 E(I<br />

H(E) =<br />

p − E † E) −1/2<br />

E † (I p − EE † ) −1/2 (I p − E † E) −1/2 . (5.48)<br />

It holds that H(E) is Hermitian, J-unitary and invertible with inverse H(E) −1 = H(−E).<br />

Also, the matrix function S u,w (z) is defined by:<br />

( ( ) )<br />

z−w<br />

I p +<br />

S u,w (z) :=<br />

1−wz − 1 uu † 0<br />

. (5.49)<br />

0 I p<br />

For lossless FIR systems, the special case w = 0 is important. In that case:<br />

( )<br />

u Ip − uu †<br />

Û =<br />

0 u † , (5.50)<br />

(<br />

v I p − (1 − √ )<br />

1 − ‖v‖<br />

̂V =<br />

2 )<br />

√ vv†<br />

‖v‖ 2<br />

. (5.51)<br />

1 − ‖v‖<br />

2<br />

−v †<br />

In the following subsection we will discuss the use of the tangential Schur algorithm<br />

to parameterize scalar wavelets. It is shown how to recover the product formula (5.7) by<br />

making suitable choices in the algorithm.<br />

5.3.2 Parameterization of scalar wavelets<br />

In Section 5.2 a lattice filter implementation of 2 × 2 lossless systems was presented. The<br />

associated parameterization procedure employs the product formula (5.7), which can be<br />

generated through a recursion which involves premultiplication in each iteration step:<br />

G (k) (z) = R k Λ(z)G (k−1) (z) (5.52)<br />

G (1) (z) = R 1 Λ(−1). (5.53)


96 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

Here, the function G (k) (z) denotes a lossless function of degree k−1, parameterized by the<br />

k parameters θ 1 , θ 2 , . . . , θ k . When attempting to rewrite this in terms of the LFTs and the<br />

tangential Schur algorithm of the previous subsection, it is noted that postmultiplication<br />

is more conveniently represented in that formalism than premultiplication. For that<br />

reason, we address the transposed expressions, which are given by:<br />

(<br />

T (<br />

T<br />

G (z)) (k) = G (z)) (k−1) Λ(z)R<br />

T<br />

k (5.54)<br />

(<br />

T<br />

G (z)) (1) = Λ(−1)R1 T . (5.55)<br />

From these equations it is not difficult to read off a valid interpolation condition and to<br />

construct an associated J-inner matrix Θ (k) (z) which allows one to rewrite an iteration<br />

step in terms of an LFT:<br />

(<br />

( T ( ) ) T<br />

G (z)) (k) = TΘ (k) (z) G (k−1) (z) , (5.56)<br />

for<br />

Θ (k) (z) =<br />

( (Λ(z)R ) )<br />

T −1 ( )<br />

k<br />

0 Rk Λ(z −1 ) 0<br />

=<br />

0 I 2 0 I 2<br />

(5.57)<br />

The following choices can then be made for the parameters in (5.43), with the recursion<br />

index k running from k = 2 to k = n, to recover (5.7) for the scalar wavelet<br />

case:<br />

u k =<br />

( ) sin(θk )<br />

− cos(θ k )<br />

(5.58)<br />

v k =<br />

( 0<br />

0)<br />

(5.59)<br />

w k = 0 (5.60)<br />

ξ k = 1 (5.61)<br />

⎛<br />

⎞<br />

cos(θ k ) − sin(θ k ) 0 0<br />

M k = ⎜sin(θ k ) cos(θ k ) 0 0<br />

⎟<br />

⎝ 0 0 1 0⎠ . (5.62)<br />

0 0 0 1<br />

M k is a block diagonal matrix with the unitary<br />

(<br />

matrices P k = R k and Q k = I 2 on<br />

0<br />

its diagonals. Choosing w k = 0 and v k = in each recursion step, implies that<br />

0)<br />

(<br />

G (k) (z) ) T<br />

has all its poles at z = 0, thus ensuring the FIR property for this polyphase<br />

filter. (This corresponds to a Potapov decomposition of ( G (k) (z) ) T<br />

.) The parameter θk<br />

shows up in the normalized direction vector u k , which gives a non-standard choice for<br />

parameterizing lossless functions. The standard choice to parameterize lossless functions<br />

requires u k , w k , ξ k and M k to be fixed and v k to contain the free parameters subject to


5.3. MULTIWAVELET PARAMETERIZATION AND DESIGN 97<br />

the constraint ‖v k ‖ < 1. The current non-standard choice, where v k is kept fixed and<br />

u k is varied, does produce a parameterization however which works properly, although<br />

some lossless functions can be obtained for different choices of parameters.<br />

It is now straightforward to verify that the LFT in each recursion step admits a<br />

representation as a mapping F U,V . Referring to Theorem 5.3.7 we have that:<br />

⎛ √ ( √ ) ⎞<br />

1−|wk |<br />

√ 2<br />

⎜<br />

u 1−|wk | 2 ‖v k ‖ 2 k I p − 1 + w k 1−‖vk ‖<br />

√ 2<br />

u k u †<br />

1−|wk | 2 ‖v k ‖ 2 k⎟<br />

Û =<br />

=<br />

̂V =<br />

=<br />

⎝<br />

√ √<br />

w k 1−‖vk ‖<br />

√ 2<br />

1−|wk |<br />

√ 2<br />

1−|wk | 2 ‖v k ‖ 2 1−|wk | 2 ‖v k ‖ u† 2 k<br />

⎛<br />

sin(θ k ) cos 2 ⎞<br />

(θ k ) sin(θ k ) cos(θ k )<br />

⎝− cos(θ k ) sin(θ k ) cos(θ k ) sin 2 (θ k ) ⎠ . (5.63)<br />

0 sin(θ k ) − cos(θ k )<br />

⎛ √ ( √ ) ⎞<br />

1−|wk |<br />

√ 2<br />

⎜<br />

1−‖vk ‖<br />

1−|wk |<br />

⎝<br />

‖v k ‖ 2 k I p − 1 + √ 2 v k v † k<br />

1−|wk | 2 ‖v k ‖ 2 ‖v k ‖<br />

√ 2<br />

⎟<br />

⎠<br />

1−‖vk ‖<br />

√ 2<br />

− 1−|w k |<br />

√ 2<br />

1−|wk | 2 ‖v k ‖ 2 1−|wk | 2 ‖v k ‖ v† 2 k<br />

⎛ ⎞<br />

0 1 0<br />

⎝0 0 1⎠ . (5.64)<br />

1 0 0<br />

The matrices U and V that are used in the mappings F U,V in the steps of the<br />

corresponding state-space recursion (5.37) are given by:<br />

⎛<br />

⎞ ⎛<br />

⎞<br />

1 0 0<br />

sin(θ k ) cos(θ k ) 0<br />

U = Û ⎜<br />

⎟ ⎜<br />

⎟<br />

⎝ 0<br />

⎠ = ⎝ − cos(θ k ) sin(θ k ) 0 ⎠ (5.65)<br />

P k<br />

0<br />

0 0 −1<br />

⎛<br />

⎞ ⎛ ⎞<br />

1 0 0<br />

0 1 0<br />

V = ̂V ⎜<br />

⎟ ⎜ ⎟<br />

⎝ 0<br />

⎠ = ⎝ 0 0 1 ⎠ . (5.66)<br />

Q k<br />

0<br />

1 0 0<br />

The resulting state-space recursion (5.37) then attains the following form:<br />

(<br />

)<br />

B (k) A (k)<br />

D (k) C (k)<br />

⎛<br />

cos(θ k )D (k−1)<br />

1,1 sin(θ k )D (k−1)<br />

1,1 −D (k−1)<br />

1,2<br />

C (k−1)<br />

1,1 . . . C (k−1)<br />

1,k−1<br />

cos(θ k )D (k−1)<br />

2,1 sin(θ k )D (k−1)<br />

2,1 −D (k−1)<br />

2,2<br />

=<br />

⎜<br />

⎝<br />

.<br />

cos(θ k )B (k−1)<br />

k−1,1<br />

⎠<br />

C (k−1)<br />

2,1 . . . C (k−1)<br />

2,k−1<br />

sin(θ k ) − cos(θ k ) 0 0 . . . 0<br />

cos(θ k )B (k−1)<br />

1,1 sin(θ k )B (k−1)<br />

1,1<br />

.<br />

sin(θ k )B (k−1)<br />

k−1,1<br />

−B (k−1)<br />

1,2<br />

.<br />

−B (k−1)<br />

k−1,2<br />

A (k−1)<br />

⎞<br />

⎟<br />

⎠<br />

(5.67)


98 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

From the previous equation it can be seen that the dynamical matrix A (k−1) is extended<br />

by means of a zero row and a first column that is equal to the second column of<br />

B (k−1) . Hence A (k) is strictly lower triangular, having a zero diagonal. Consequently its<br />

eigenvalues are zero and the system is FIR.<br />

Note that (5.67) yields a balanced realization of G (k) (z) T and it may be transposed<br />

in order to obtain a balanced realization of G (k) (z). It follows that A (k)T is a strictly<br />

upper triangular matrix: it is in real Schur form. (Balancing leaves the freedom of the<br />

orthogonal group which allowed us to bring A (k)T into that real Schur form.) Since the<br />

real Schur form is not necessarily unique it is not a canonical form; this is a consequence<br />

of the non-standard parameterization mentioned earlier.<br />

5.3.3 Parameterization of multiwavelets<br />

The parameterization in the previous section can be extended to the multiwavelet case.<br />

For the parameters of the elementary J-inner factors from (5.43) the following choices<br />

are made: Since we are building FIR filters we take w = 0 and v = (0 . . . 0) T . The<br />

parameter ξ can be chosen freely on the unit circle; the choice ξ = 1 is made so that<br />

S u,w (ξ) = I 4r . In order to again ensure that T Θ(u,v,w,ξ,M)(z) coincides with a mapping<br />

F U,V as in Theorem 5.3.7, due to the convenient choice of ξ we need to have that M =<br />

( ) P 0<br />

.<br />

0 Q<br />

Theorem 5.3.8. Let G (k+1) (z) be a FIR all-pass filter of order k + 1. Then with the<br />

choice of w k = 0, v k = (0 . . . 0) T , ‖u k ‖ = 1, ξ k = 1, P and Q unitary, it follows that<br />

G (k+1) (z) can be factored as:<br />

G (k+1) (z) = (I 2r + (z −1 − 1)u k u † k )P kG (k) (z)Q † k . (5.68)<br />

where G (k) (z) is a FIR all-pass filter of order k.<br />

Proof. Obtaining a state-space recursion for a block diagonal matrix G k involves a factorization<br />

of Θ (k) (z) as in [55, Proposition 5.2]:<br />

Θ (k) (u k , v k , w k , ξ k , M k )(z) = H(u k v † k )S u k ,w k<br />

(z)S uk ,w k<br />

(ξ k ) −1 H(u k v † k )−1 M k . (5.69)<br />

Also observe that for the choices made [55, Theorem 6.4] holds, such that:<br />

( )<br />

Θ (k) (u k , v k , w k , ξ k , M k )(z) = ̂Θ (k) Pk 0<br />

(u k , v k , w k )(z)<br />

, (5.70)<br />

0 Q k<br />

with<br />

̂Θ (k) (u k , v k , w k )(z) = S uk ,0(z) =<br />

(<br />

)<br />

I 2r + (z − 1)u k u † k<br />

0<br />

. (5.71)<br />

0 I 2r<br />

Using the LFT formula in (5.38) we see that the inverse of the upper right quadrant of<br />

(5.71) is used as a post multiplying factor. However we are working on the transposed


5.3. MULTIWAVELET PARAMETERIZATION AND DESIGN 99<br />

system as noted in (5.56), such that after transposition the inverse factor appears as a<br />

pre-multiplying factor. This leads to the recursion formula:<br />

( )<br />

G (k+1) (z) T = T Θ(uk ,v k ,w k ,ξ k ,M k )(z) G (k) (z) T = Q k G (k) (z) T P † k (I 2r + (z −1 − 1)u k u † k ),<br />

(5.72)<br />

such that:<br />

G (k+1) (z) = (I 2r + (z −1 − 1)u k u † k )P kG (k) (z)Q † k<br />

(5.73)<br />

Observe that this is a more general parameterization than previously available in the<br />

literature. One can obtain a lattice decomposition in terms of Givens rotation matrices<br />

from Theorem 5.3.8 with the following choices:<br />

We relate the unitary matrix P to the vector u, which will again contain the parameters.<br />

Analogous to the scalar case we require that P † u = −e 2r and any such P will do.<br />

Here e n denotes the n th standard basis vector. As in the scalar case we choose Q = I 2r .<br />

For the parameter vector u it must again hold that ‖u‖ = 1 and as a choice for a<br />

fully parameterized u one can for example make:<br />

⎛<br />

⎞<br />

sin(α 1 ) sin(α 2 ) sin(α 3 )<br />

u = ⎜− cos(α 3 ) sin(α 1 ) sin(α 2 )<br />

⎟<br />

⎝ cos(α 2 ) sin(α 1 ) ⎠ . (5.74)<br />

− cos(α 1 )<br />

Corollary 5.3.9. Let G (k+1) (z) be a FIR all-pass filter of order k + 1. Then with the<br />

choice of w k = 0, v k = (0 . . . 0) T , ‖u k ‖ = 1, ξ k = 1, Q = I 2r and P † k u k = −e 2r it follows<br />

that G (k+1) (z) can be factored in a lattice decomposition as:<br />

⎛<br />

⎞<br />

1<br />

.<br />

G (k+1) .. (z) = P k ⎜<br />

⎟<br />

⎝ 1 ⎠ G(k) (z), (5.75)<br />

z −1<br />

where G (k) (z) is a FIR all-pass filter of order k.<br />

Proof. Observe that due to the fact that P k is a unitary 2r × 2r matrix and that P k u k =<br />

−e 2r , the following equalities hold:<br />

Such that (5.68) can be rewritten as:<br />

u k = −P k e 2r , (5.76)<br />

u † k P k = −e † 2r P † k P k = −e 2r , (5.77)<br />

G (k+1) (z) = P k (I 2r + (z −1 − 1)e 2r e † 2r )G(k) (z)Q † k<br />

(5.78)<br />

⎛<br />

1<br />

. .. = P k ⎜<br />

⎟<br />

⎝ 1 ⎠ G(k) (z)Q † k . (5.79)<br />

z −1 ⎞


100 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

Since Q k = I 2r the following product formula is obtained:<br />

⎛<br />

⎞<br />

1<br />

.<br />

G (k+1) .. (z) = P k ⎜<br />

⎟<br />

⎝ 1 ⎠ G(k) (z), (5.80)<br />

z −1<br />

which completes the proof.<br />

A practical choice for P relating to the choice of u as in (5.74) is to build it as a<br />

product of Givens rotation matrices as in for example [116]. Such a parameterization for<br />

r = 2 is given by:<br />

⎛<br />

⎞ ⎛<br />

⎞<br />

cos(α 3 ) − sin(α 3 ) 0 0 1 0 0 0<br />

P (α 1 , α 2 , α 3 ) = ⎜sin(α 3 ) cos(α 3 ) 0 0<br />

⎟ ⎜0 cos(α 2 ) − sin(α 2 ) 0<br />

⎟<br />

⎝ 0 0 1 0⎠<br />

⎝0 sin(α 2 ) cos(α 2 ) 0⎠<br />

0 0 0 1 0 0 0 1<br />

⎛<br />

⎞<br />

1 0 0 0<br />

⎜0 1 0 0<br />

⎟<br />

⎝0 0 cos(α 1 ) − sin(α 1 ) ⎠<br />

0 0 sin(α 1 ) cos(α 1 )<br />

This results in the following parameterized matrix P :<br />

P =<br />

( cos(α3) − cos(α 2) sin(α 3) cos(α 1) sin(α 2) sin(α 3) − sin(α 1) sin(α 2) sin(α 3)<br />

sin(α 3) cos(α 2) cos(α 3) − cos(α 1) cos(α 3) sin(α 2) cos(α 3) sin(α 1) sin(α 2)<br />

0 sin(α 2) cos(α 1) cos(α 2) − cos(α 2) sin(α 1)<br />

0 0 sin(α 1) cos(α 1)<br />

)<br />

, (5.81)<br />

which indeed satisfies the condition P † k u k = −e 2r .<br />

The parameterization as in the so-called “great factorization theorem”, cf. [107, 117],<br />

can be obtained with the choice of P k = Q k = I 2r .<br />

Corollary 5.3.10. Let G (k+1) (z) be a FIR all-pass filter of order k + 1. Then with the<br />

choice of w k = 0, v k = (0 . . . 0) T , ‖u k ‖ = 1, ξ k = 1, Q = I 2r and P = I 2r it follows that<br />

G (k+1) (z) can be factored in a Householder decomposition as:<br />

(<br />

G (k+1) (z) = I 2r − u k u k + u k u † k z−1) G (k) (z), (5.82)<br />

where G (k) (z) is a FIR all-pass filter of order k.<br />

Proof. Consider (5.68) from Theorem 5.3.8. With the choice of P k = Q k = I 2r , (5.68)<br />

can be rewritten as:<br />

(<br />

G (k+1) (z) = I 2r − u k u k + u k u † k z−1) G (k) (z). (5.83)


5.3. MULTIWAVELET PARAMETERIZATION AND DESIGN 101<br />

Using the parameterization in Corollary 5.3.9 of multiwavelets as polyphase all-pass<br />

systems all requirements but the first (balanced) vanishing moment are enforced. In the<br />

next section it is discussed how a first balanced vanishing moment can be built in as<br />

first discussed in [95]. It will turn out that the first balanced vanishing moment can be<br />

enforced by initializing the recursion in a specific way. In Section 5.3.5 a step-by-step<br />

design procedure for orthogonal multiwavelets will be provided.<br />

5.3.4 Balanced vanishing moments<br />

When considering a constant signal, we may investigate how it is processed when employing<br />

Mallat’s cascade algorithm [79]. For scalar wavelets, the approximation coefficients<br />

will all be identical, i.e., a constant coefficient vector. Therefore the sampled signal is<br />

commonly used to initialize the multi resolution analysis, because the sampled signal<br />

and the approximation coefficients constitute similar sequences. This avoids an actual<br />

computation of the scaling coefficients that express the input signal in terms of the basis<br />

spanned by shifted versions of the scaling function to initialize the multi resolution<br />

analysis and it also avoids the actual computation of the multiscaling and multiwavelet<br />

function.<br />

When processing a constant signal in the multiwavelet case, one finds that each subsequence<br />

of scaling coefficients, for each of the multiwavelets, is also constant. However,<br />

when these constants differ per multiwavelet, then one cannot simply use the sampled signal<br />

to initialize the multi resolution analysis. The necessary condition to use a sampled<br />

signal to initialize the multiresolution analysis is that the low-pass synthesis operator<br />

should exactly preserve polynomials of order p for a p th order balanced vanishing moment<br />

[77]. If this condition were not satisfied a preprocessing step would be required,<br />

which is unattractive. The alternative is to impose a so-called balancing condition on<br />

the multiwavelet structure and to allow again the direct use of the sampled signal.<br />

A 0 th order balanced scaling function preserves constant signals [27]:<br />

∫<br />

R<br />

∫<br />

1φ [0]<br />

(t)dt = . . . =<br />

R<br />

1φ [r−1]<br />

(t)dt. (5.84)<br />

Due to power complementarity this is equivalent to the condition that the detail coefficients<br />

are not affected by constant signals: 0 = ∫ R 1ψ (t)dt. When taking the dilation<br />

[k]<br />

equation (3.24) and wavelet equation (3.25) and integrating both sides over the real axis<br />

we obtain:<br />

ζ 0 =<br />

η 0 =<br />

∫<br />

∫<br />

R<br />

R<br />

φ(t)dt = √ 2n−1<br />

∑<br />

∫<br />

2 C k<br />

k=0<br />

ψ(t)dt = √ 2n−1<br />

∑<br />

∫<br />

2 D k<br />

k=0<br />

R<br />

R<br />

φ(2t + k)dt, (5.85)<br />

φ(2t + k)dt. (5.86)<br />

The desired vanishing moment comes down to the condition η 0 = 0. The balancing


102 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

⎛ ⎞<br />

1<br />

⎜<br />

condition comes down to ζ 0 = c<br />

. ⎟<br />

⎝.<br />

⎠, for some constant scalar c. Using the change of<br />

1<br />

basis τ = 2t + k, dτ = 2dt and dt = 1 2dτ we obtain:<br />

∫<br />

∫<br />

φ(2t + k)dt = φ(τ) 1 2 dτ = 1 2 ζ 0, (5.87)<br />

from which it follows that:<br />

R<br />

R<br />

ζ 0 = 1 2√<br />

2H0 (1)ζ 0 , (5.88)<br />

η 0 = 1 2√<br />

2H1 (1)ζ 0 . (5.89)<br />

This can be rewritten in terms of the polyphase matrix H p (z) as:<br />

( )<br />

ζ0<br />

= 1 ( )<br />

√ ζ0<br />

2Hp (1) . (5.90)<br />

η 0 2 ζ 0<br />

Since H p (z) is all-pass and thus orthogonal at z = 1 we have that:<br />

( )<br />

1<br />

2 ‖H ζ0<br />

p(1) ‖ 2 = 1 ( )<br />

ζ 0 2 ‖ ζ0<br />

‖ 2 = ‖ζ 0 ‖ 2 , (5.91)<br />

ζ 0<br />

( )<br />

ζ0<br />

‖ ‖ 2 = ‖ζ 0 ‖ 2 + ‖η 0 ‖ 2 , (5.92)<br />

η 0<br />

and since these expressions are equal due to (5.90) it follows that‖η 0 ‖ 2 = 0 and η 0 = 0.<br />

It can then be concluded that the first vanishing moment is built-in, meaning that every<br />

feasible (multi)wavelet structure which fits the orthogonal set-up as discussed previously<br />

in this chapter, already obeys the conditions for a first vanishing moment. The conclusion<br />

here is that the first vanishing moment is implicitly required and restricts the class of<br />

lossless polyphase matrices H 0 (z), H 1 (z). The conditions for a first balanced vanishing<br />

moment can be enforced in the tangential Schur algorithm by means of an interpolation<br />

condition on the unit circle. The condition on H p<br />

(k) (z) for the first balanced vanishing<br />

moment is as in [95]:<br />

Theorem 5.3.11. If H p<br />

(k) (z) is a 2r×2r real FIR all-pass filter of order k associated with<br />

the corresponding FIR filters H 0 (z), H 1 (z) and vector functions φ(t) and ψ(t), satisfying<br />

the dilation an wavelet equation, and √ 2 is a simple eigenvalue of H 0 (1), then ψ(t) has<br />

a balanced vanishing moment of order 0 if and only if<br />

(1, 1, . . . , 1|1, 1, . . . , 1) H (k)<br />

p (1) T = √ 2 (1, 1, . . . , 1|0, 0, . . . , 0) (5.93)<br />

Proof. This theorem directly follows from (5.90). Note that the condition that √ 2 is<br />

a simple eigenvalue of H 0 (1) ensures that (5.88) has a solution ζ 0 which is determined<br />

up to a nonzero scaling factor. Orthogonality of the multiresolution structure and sign<br />

conventions for φ(t) and ψ(t) fix this scaling factor.


5.3. MULTIWAVELET PARAMETERIZATION AND DESIGN 103<br />

The vanishing moment condition is now rewritten in terms of H 0 (z), H 1 (z). The<br />

following relation holds:<br />

( ) ( )<br />

1√ 2Hp (z 2 Ir 0 Ir I<br />

) r<br />

2<br />

0 z −1 = 1 ( )<br />

√ H0 (z) H 2 0 (−z)<br />

. (5.94)<br />

I r I r −I r 2 H 1 (z) H 1 (−z)<br />

This equation allows one to relate the derivatives of H p (z) to the derivatives of H 0 (z),<br />

H 1 (z), H 0 (−z) and H 1 (−z), in particular at z = 1. Additionally, due to orthogonality,<br />

we have that H 1 (1) = 0, H 1(1) ′ = 0, H 1 ”(1) = 0 is equivalent to H 0 (−1) = 0, H 0(−1) ′ =<br />

0, H 0 ”(−1) = 0. This specifies the vanishing moment condition entirely in terms of the<br />

low-pass filter. Note that all matrices on the left hand side of (5.94) are real lossless<br />

matrices, hence the matrix on the right hand side is real lossless and thus orthogonal at<br />

|z| = 1 and more specifically at z = 1:<br />

(<br />

1√ H0 (1) 2<br />

2 H 1 (1)<br />

)<br />

H 0 (−1)<br />

H 1 (−1)<br />

(<br />

1√ Ir<br />

= H p (1)I 2r 2<br />

2<br />

)<br />

I r<br />

I r −I r<br />

. (5.95)<br />

Using the right hand side of (5.95) and that η 0 = 0, (5.90) can be rewritten as:<br />

( )<br />

ζ0<br />

= H p (1) 1 ( ) ( )<br />

√ Ir I 2 r ζ0<br />

0 2 I r −I r 0<br />

= 1 2<br />

√<br />

2<br />

(<br />

H0 (1)<br />

H 0 (−1)<br />

H 1 (1) H 1 (−1)) (<br />

ζ0<br />

0<br />

(5.96)<br />

)<br />

. (5.97)<br />

One can investigate how this condition translates to H p (1) in the scalar case, i.e.,<br />

√<br />

r = 1. The conditions become: 2ζ0 = H 0 (1)ζ 0 and 0 = H 1 (1)ζ 0 . Since ζ 0 ≠ 0, it<br />

must hold that H 0 (1) = √ 2 and H 1 (1) = 0 and due to power complementarity that<br />

H 0 (−1) = 0 and thus that H 1 (−1) = ± √ 2. The sign choice for H 1 (−1) corresponds to<br />

a sign choice for the wavelet ψ(t). We shall use the convention H 1 (−1) = √ 2. Hence<br />

( )<br />

√ 1 1<br />

H p (1) = 1 2 2 . In the multiwavelet case balancing is required and we have to<br />

1 −1<br />

work with the scalar multiples.<br />

Using (5.97), the condition for the first (0 th order) vanishing moment becomes [77]:<br />

√<br />

T<br />

2 (1, 1, . . . , 1) = H 0 (1) (1, 1, . . . , 1) T (5.98)<br />

(0, 0, . . . , 0) T = H 1 (1) (1, 1, . . . , 1) T (5.99)<br />

As a side remark it can be verified that the results in this section are consistent with<br />

the condition for a first balanced vanishing moment in [77, Theorem 2]. Taking into<br />

account the differences in notation and terminology the balanced vanishing moment is<br />

characterized in [77] by the conditions:<br />

(1, 1, . . . , 1) H 0 (1) = √ 2 (1, 1, . . . , 1) (5.100)<br />

(1, 1, . . . , 1) H 0 (−1) = √ 2 (0, 0, . . . , 0) (5.101)<br />

The conditions in (5.100) and (5.101) relate to the conditions in (5.98) and (5.99) due to<br />

the following lemma, which is a application to (5.93):


104 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

Lemma 5.3.12. Consider an orthogonal real matrix N with a real eigenvalue λ and a<br />

corresponding real eigenvector v. Then the equality Nv = λv is equivalent to v T N = λv T .<br />

Proof. Consider the following equality:<br />

The norm on both sides needs to be equal as well:<br />

Nv = λv (5.102)<br />

v T N T Nv = |λ| 2 v T v, (5.103)<br />

v T v = |λ| 2 v T v, (5.104)<br />

hence |λ| 2 = 1 and thus for real eigenvalues we have λ = 1 or λ = −1. Now construct an<br />

orthogonal matrix N 2 which has v as its first column: v = N 2 e 1 . We can then rewrite<br />

(5.102) as:<br />

NN 2 e 1 = λN 2 e 1 (5.105)<br />

N T 2 NN 2 e 1 = λe 1 (5.106)<br />

Due to the fact that |λ| = 1, the orthogonal matrix N2 T NN 2 will have the following<br />

structure:<br />

⎛<br />

⎞<br />

λ 0 . . . 0<br />

0<br />

N2 T NN 2 =<br />

⎜<br />

⎟<br />

⎝ . N3 ⎠ ,<br />

0<br />

where N 3 is another orthogonal matrix. Hence we can write:<br />

which completes the proof.<br />

e T 1 N T 2 NN 2 = λe T 1 , (5.107)<br />

e T 1 N T 2 N = λe T 1 N T 2 , (5.108)<br />

v T N = λv T , (5.109)<br />

A consequence of this lemma is that the transpose in (5.93) can be omitted.<br />

The fact that the parameterization in Corollary 5.3.9 has the convenient property<br />

that H p<br />

(k+1) (1) = H p<br />

(k) (1), makes it possible to build in a vanishing moment with respect<br />

to the multiwavelet in a straightforward manner.<br />

Theorem 5.3.13. An orthogonal multiwavelet of multiplicity r having a balanced vanishing<br />

moment and a FIR filter of order k can be built as a polyphase all-pass H p<br />

(k) (z)<br />

by initializing the factorization in Corollary 5.3.9 with a zeroth order polyphase all-pass<br />

H p<br />

(0) that is factored as:<br />

( ) 1 0<br />

H p (0)T = H α H β , (5.110)<br />

0 L


5.3. MULTIWAVELET PARAMETERIZATION AND DESIGN 105<br />

where H α and H β are fixed orthogonal Householder matrices of the form:<br />

with:<br />

H α = I 2r − 2 ααT<br />

α T α,<br />

(5.111)<br />

α T = (1, 1, . . . , 1|0, 0, . . . , 0) / √ r − (1, 0, . . . , 0|0, 0, . . . , 0) (5.112)<br />

β T = (1, 1, . . . , 1|1, 1, . . . , 1) / √ 2r − (1, 0, . . . , 0|0, 0, . . . , 0) , (5.113)<br />

and L is an arbitrary (2r − 1) × (2r − 1) orthogonal matrix.<br />

Proof. First observe that α and β are of the form α =<br />

a−b<br />

‖a−b‖. The Householder reflection<br />

H α x reflects the component of x in the direction a to b. Components that are orthogonal<br />

to a are left unchanged.<br />

Since (5.71) shows that ̂Θ (k) (u k , v k , w k )(1) = I 2r , ∀k interpolation conditions on<br />

H p<br />

(k) (1) that are built-in in H 0 are preserved throughout the recursion. It is now left to<br />

correctly initialize this recursion.<br />

The interpolation condition of interest is the balancing condition from Theorem 5.3.11:<br />

H (0)<br />

p (1, 1, . . . , 1|1, 1, . . . , 1) T = (1, 1, . . . , 1|0, 0, . . . , 0) T . (5.114)<br />

Now suppose that we factor H p (0) as the product of three matrices H p<br />

(0) = W RU, with<br />

W and U symmetrical. Then we get the condition:<br />

W RU (1, 1, . . . , 1|1, 1, . . . , 1) T = (1, 1, . . . , 1|0, 0, . . . , 0) T , (5.115)<br />

which can be rewritten as (using the symmetry property):<br />

RU (1, 1, . . . , 1|1, 1, . . . , 1) T = W (1, 1, . . . , 1|0, 0, . . . , 0) T . (5.116)<br />

If the matrix W is chosen to be H α then the right hand side of (5.116) simplifies to e 1<br />

and similarly if U is chosen to be H β then the left hand side of (5.116) simplifies to Re 1<br />

as a direct consequence of the theory on Householder transformations. Thus with this<br />

choice of W and U any matrix of the following form satisfies the condition in (5.116):<br />

( ) 1 0<br />

. (5.117)<br />

0 L<br />

Since we want to obtain an orthogonal system the initial matrix H p<br />

(0) and thus L has<br />

to be orthogonal. Since it is multiplied with orthogonal matrices during the recursion<br />

orthogonality will be preserved.<br />

Note that such an orthogonal matrix L can be parameterized by (2 − 1)(2r − 1) free<br />

angular parameters in a straightforward way. For example for r = 2 a 3 × 3 matrix Q<br />

can be parameterized as:<br />

( ) ( 1 0 I2 0<br />

=<br />

0 L 0 R(θ 3 )<br />

) ⎛ ⎞<br />

0 0<br />

⎝1<br />

0 R(θ 2 ) 0⎠<br />

0 0 1<br />

( )<br />

I2 0<br />

. (5.118)<br />

0 R(θ 1 )


106 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

( )<br />

cos(θ) − sin(θ)<br />

R(θ) =<br />

sin(θ) cos(θ)<br />

(5.119)<br />

5.3.5 Multiwavelet design<br />

An explicit algorithm for constructing a multiwavelet filter of order n with r = 2 is given<br />

by:<br />

1. Initialize the recursion:<br />

a) Select initial angles θ1, 0 θ2, 0 θ3.<br />

0<br />

( ) 1 0<br />

b) Construct the matrix from (5.118) using parameters θ 0<br />

0 L<br />

1, θ2, 0 θ3.<br />

0<br />

c) Construct the matrices U and W from (5.116) as:<br />

U =<br />

W =<br />

⎛<br />

⎜<br />

⎝<br />

1 1<br />

2 2<br />

1 1<br />

2<br />

1<br />

2<br />

− 1 2<br />

1<br />

d) Construct the zeroth order all-pass H (0)<br />

p<br />

2. Set k = 1 and enter the recursion<br />

3. Recursion step<br />

a) Select angles θ k 1, θ k 2, θ k 3.<br />

1<br />

2<br />

1<br />

2<br />

2<br />

− 1 2<br />

− 1 2⎟<br />

1 ⎠ (5.120)<br />

2<br />

− 1 2<br />

1<br />

2<br />

⎞<br />

2<br />

− 1 2<br />

− 1 2<br />

⎛ √<br />

1<br />

2 2<br />

1 ⎞<br />

2√<br />

2 0 0<br />

√ 1<br />

⎜ 2 2 −<br />

1<br />

2√<br />

2 0 0<br />

⎟<br />

⎝ 0 0 1 0⎠ (5.121)<br />

0 0 0 1<br />

as:<br />

H (0)<br />

p = W RU. (5.122)<br />

b) Use these angles to construct the matrix P k as in (5.81).<br />

c) Recursively construct a k th order multiwavelet filter H p<br />

(k) (z) from a (k − 1) th<br />

order filter in accordance to Corollary 5.3.9:<br />

( )<br />

H p (k) I3 0<br />

= P k<br />

0 z −1 H p (k−1)<br />

(5.123)<br />

4. If k < n then set k = k + 1 and goto 3.<br />

The parameterization of multiwavelets as stable all-pass systems in the previous section<br />

allows for the design of multiwavelets for a specific purpose. The discussion on a<br />

suitable criterion for wavelet design as in Section 5.1 applies here as well since energy<br />

preservation holds when considering the overall multiwavelet filter.


5.3. MULTIWAVELET PARAMETERIZATION AND DESIGN 107<br />

Wavelet coefficients 1<br />

A_6<br />

D_6<br />

200<br />

D_5<br />

D_4<br />

0<br />

D_3<br />

D_2<br />

D_1 -200<br />

83 120 160 230<br />

Wavelet coefficients 2<br />

A_6<br />

400<br />

D_6<br />

D_5<br />

200<br />

D_4<br />

0<br />

D_3<br />

D_2<br />

-200<br />

D_1 -400<br />

83 120 160 230<br />

Prototype signal<br />

400<br />

200<br />

0<br />

83 120 160 230<br />

Figure 5.11: Windowing for multiwavelet design<br />

The orthogonality between the different filters in multiwavelets make them conceptually<br />

an appealing tool to distinguish and detect features simultaneously. In that respect<br />

it is opportunistic to employ r time-scale masks (as in Figure 5.11) to the wavelet coefficients.<br />

So for a multiwavelet with r = 2 two time-scale masks are used such that two<br />

features can be discriminated. When using these masks for optimization the conservation<br />

of energy assumption no longer holds. Also it is not convenient to use a criterion (such<br />

as l 1 minimization) that involves minimization since it is possible that the optimization<br />

routine converges to an optimum that is an effective delay such that features are pushed<br />

outside the mask. It is still possible to use the l 4 maximization, however since we want<br />

to put the energy at a specific place it is also possible to use l 2 maximization. Another<br />

point of interest is that for detection purposes it is most convenient to have full resolution<br />

at each scale, i.e. to use an undecimated multiwavelet transform.<br />

As was the case for scalar wavelet design the optimization is performed by means of<br />

a local search technique with random starting points. The choice of parameterization<br />

ensures that all required conditions in order to obtain an orthogonal stable all-pass with<br />

at least one vanishing moment is built-in. However for applications such as filtering


108 CHAPTER 5. ORTHOGONAL WAVELET DESIGN<br />

additional smoothness might be needed. For the smoothness of multiwavelet functions a<br />

vanishing moment is not sufficient. Additionally the multiwavelet needs to be balanced<br />

[76, 77, 27]. Balancing requires an additional interpolation condition that is difficult<br />

to handle in the current parameterization and is still an open problem. It is still not<br />

guaranteed that the resulting system is indeed a multiwavelet since it is not enforced<br />

that each set of wavelet coefficients forms a high-pass filter and the associated scaling<br />

coefficients form a low-pass filter.


Chapter 6<br />

Biomedical applications of orthogonal<br />

(multi)wavelet design<br />

In this chapter the potential of the approaches, as introduced in Chapter 5, will be<br />

demonstrated by means of a number of practical applications. In Section 6.1 two practical<br />

applications of (multi)wavelet design with respect to cardiac signal processing are<br />

discussed. In Section 6.1.1 it is demonstrated how scalar wavelet design is useful for the<br />

detection of QRS complexes in ECG signals. In Section 6.1.2 the length of the QT interval<br />

is estimated using designed multiwavelets. In Section 6.2 the use of wavelet design for<br />

an application in 2D image analysis is discussed. This describes the removal an artifact<br />

in magnetic resonance imaging, called the “bias field”, using specially designed wavelets<br />

[64].<br />

6.1 Applications of orthogonal (multi)wavelet design in cardiology<br />

6.1.1 Detecting the QRS complex using orthogonal wavelet design<br />

Using the wavelet design approach in Chapter 5 it is possible to design a wavelet for<br />

a given signal of interest. Given an ECG signal (see Section 2.3) one can average the<br />

beats in the signal and use them as a prototype for the wavelet design algorithm. If<br />

an ECG signal consisting of multiple beats was used as a prototype, the wavelet design<br />

would have searched for a wavelet that performs well for a superposition of the beats<br />

in the ECG signal. Hence it is justified to construct a prototype by aligning the ECG<br />

beats. The annotated R peak is used as a reference for the alignment. Since the QRS<br />

complex is the dominant part of an ECG beat, the location of this complex is generally<br />

not hard to detect. However as discussed in Section 2.2, the heart is a 3D organ. As<br />

such, the signals that are measured extracardiacally may exhibit different morphologies,<br />

dependent on the position of the lead, relative to the heart, complicating the detection.<br />

109


110 CHAPTER 6. BIOMEDICAL APPLICATIONS OF WAVELET DESIGN<br />

Daubechies 2<br />

1<br />

0.5<br />

0<br />

−0.5<br />

0 1 2 3<br />

L , 4 parameters<br />

1<br />

1<br />

0.5<br />

0<br />

−0.5<br />

0 2 4 6<br />

L , 3 parameters<br />

4<br />

1<br />

0.5<br />

0<br />

−0.5<br />

1<br />

0.5<br />

0<br />

L 1<br />

, 2 parameters<br />

0 1 2 3<br />

L 1<br />

, 5 parameters<br />

−0.5<br />

0 2 4 6 8<br />

L 4<br />

, 4 parameters<br />

1<br />

0.5<br />

0<br />

L 1<br />

, 3 parameters<br />

−0.5<br />

0 2 4<br />

4<br />

2<br />

0<br />

−2<br />

L 4<br />

, 2 parameters<br />

0 1 2 3<br />

L 4<br />

, 5 parameters<br />

1<br />

0.5<br />

0<br />

−0.5<br />

0 2 4<br />

1<br />

0<br />

−1<br />

0 2 4 6<br />

2<br />

1<br />

0<br />

−1<br />

0 2 4 6 8<br />

Figure 6.1: Wavelets used for the detection of the QRS complex in episode 103.<br />

Furthermore patient-to-patient variation and pathologies may affect the morphology of<br />

ECG signals.<br />

In order to assess whether the wavelet design procedure from Section 5.2 indeed yields<br />

a beneficial distribution of the energy in the wavelet domain for the detection of the QRS<br />

complex, the ECG beat from Figure 5.1(a) was used as a prototype signal for wavelet<br />

design for two to five parameters and for both criteria. This prototype is a superposition<br />

of the normal beats from episode 103 of the MIT-BIH arrythmia database [43]. Next<br />

the stationary wavelet transform (see Section 3.3, [91]) was calculated for 2 19 samples of<br />

episode 103 of the MIT-BIH arrythmia database. The choice of this length is a result<br />

of the used implementation of the SWT that accepts only signal with a length that is a<br />

power of two. For the detection step only a single scale was taken into account. Note that<br />

a sparse representation in the wavelet domain may yield a certain pattern in space-scale<br />

that can be used as a detector for a given morphology. A QRS complex was detected<br />

if the absolute value of the wavelet coefficients at that particular scale exceeded a fixed<br />

threshold value. This threshold value was chosen for each wavelet and each scale such that


6.1. APPLICATIONS OF WAVELET DESIGN IN CARDIOLOGY 111<br />

c 0 c 1 c 2 c 3 c 4 c 5 c 6 c 7 c 8 c 9<br />

db2 0.4830 0.8365 0.2241 -0.1294<br />

l 1 , 2p 0.5010 0.8313 0.2061 -0.1242<br />

l 1 , 3p 0.1860 0.7394 0.6336 -0.0607 -0.1125 0.0283<br />

l 1 , 4p -0.1077 0.1915 0.7849 0.5750 0.0167 -0.0668 0.0132 0.0074<br />

l 1 , 5p 0.0159 0.0279 -0.0615 0.1402 0.7140 0.6652 0.0593 -0.1380 -0.0206 0.0117<br />

l 4 , 2p 0.1039 0.7868 0.6032 -0.0797<br />

l 4 , 3p -0.1111 0.1612 0.8158 0.5443 0.0024 0.0017<br />

l 4 , 4p -0.1143 0.2268 0.8451 0.4699 -0.0190 0.0128 -0.0047 -0.0024<br />

l 4 , 5p -0.1017 0.2915 0.8666 0.3894 -0.0311 0.0023 -0.0170 0.0273 -0.0097 -0.0034<br />

Table 6.1: Low-pass filters used for the detection of the QRS complex in episode 103.<br />

The number before “p” indicates the number of parameters used to design the filter.<br />

the sum of the false negatives and false positives was minimized. Since the effectiveness of<br />

the approach is evaluated here and not the robustness of a detector, this is a valid choice<br />

for the goal at hand. If the threshold value was exceeded within 20 samples (55ms) of<br />

an annotated QRS complex this was accepted as a valid detection. If it was outside this<br />

interval a false positive was reported. This is a very basic detection scheme since we do<br />

not aim to find a detector for the QRS complex which is nearly flawless, or even tries to<br />

find the exact location of the complex, but we are assessing the potential of the wavelet<br />

design approach. Note that the false positives may be clotted together; in this case all of<br />

them are still considered individual false positives. As reference wavelet the Daubechies<br />

2 wavelet was chosen due to the fact that it is quite popular (see e.g. [105, 114, 93, 1])<br />

and that the steep slopes in the signal are captured well by the Daubechies 2 wavelet<br />

function.<br />

All designed wavelets and the Daubechies 2 wavelet were able to correctly find all<br />

or all but one of the QRS complexes at level 3 and 4 of the wavelet transform in the<br />

1690 beats of episode 103 that were considered. Upon examining the wavelet and scaling<br />

functions in Figure 6.1, we see that indeed the (possibly mirrored) Daubechies 2 wavelet<br />

was found as an optimum for some of the designed wavelets, even though in some of these<br />

cases a longer filter size was used. From the corresponding low-pass filter coefficients in<br />

Table 6.1, we see that for a large number of higher order filters a number of coefficients<br />

are close to zero and that the remaining coefficients are in the same order of magnitude<br />

as for the Daubechies 2 wavelet. This corresponds to Observation 5.2.7.<br />

Next, as a second test, a dataset was used which exhibits more variation in terms<br />

of morphology. The selected dataset: episode 215 of the MIT-BIH arrythmia database<br />

[43] does not only contain normal beats, but also atrial premature contractions and<br />

premature ventricular contractions. The normal beats as illustrated in Figure 6.2 have a<br />

different morphology than the normal beats of episode 103 as displayed in Figure 5.1(a).<br />

Of episode 215 an excerpt consisting of 2718 beats was considered. The results are<br />

displayed in Table 6.2.<br />

From Table 6.2 it can be seen that the l 1 designed wavelets perform better than the<br />

Daubechies 2 wavelet for this dataset. The results for the l 4 designed wavelets are not as<br />

good. The Daubechies 3 wavelet performs quite well. It can however be seen from the<br />

table that an increase in order for the Daubechies wavelets does not automatically yield<br />

an improvement in terms of detection performance. Among the higher order wavelets,


112 CHAPTER 6. BIOMEDICAL APPLICATIONS OF WAVELET DESIGN<br />

200<br />

150<br />

100<br />

50<br />

0<br />

−50<br />

−100<br />

50 100 150 200 250<br />

Figure 6.2: Smoothed normal beat from episode 215 of the MIT-BIH database.<br />

wavelet false neg. false pos. total errors level thres.<br />

Daubechies 2 9 3 12 4 295<br />

Daubechies 3 4 3 7 4 243<br />

Daubechies 4 20 0 20 4 311<br />

Daubechies 5 8 4 12 4 256<br />

Daubechies 6 23 5 28 4 304<br />

Daubechies 7 78 27 105 4 367<br />

Daubechies 8 71 17 88 4 347<br />

l 1 , 2 parameters 8 1 9 4 255<br />

l 1 , 4 parameters 8 3 11 4 271<br />

l 1 , 5 parameters 6 0 6 4 252<br />

l 1 , 6 parameters 4 0 4 4 240<br />

l 1 , 8 parameters 3 0 3 4 230<br />

l 4 , 2 parameters 18 0 18 4 380<br />

l 4 , 3 parameters 9 4 13 4 303<br />

l 4 , 4 parameters 8 4 12 4 294<br />

Table 6.2: Detection of the QRS complex in 2718 beats of episode 215 of the MIT-BIH<br />

arrythmia database. The used wavelet filter, the number of missed beats, the number of<br />

false detections, the total number of errors, the used wavelet scale and the used threshold<br />

value are displayed in the columns respectively.


6.1. APPLICATIONS OF WAVELET DESIGN IN CARDIOLOGY 113<br />

1<br />

0.5<br />

0<br />

Daubechies 2<br />

−0.5<br />

0 1 2 3<br />

L 1<br />

, 5 parameters<br />

1<br />

0.5<br />

0<br />

L 1<br />

, 2 parameters<br />

−0.5<br />

0 1 2 3<br />

L 1<br />

, 6 parameters<br />

1<br />

0.5<br />

0<br />

L 1<br />

, 4 parameters<br />

−0.5<br />

0 5<br />

L 1<br />

, 8 parameters<br />

1<br />

0.5<br />

0<br />

−0.5<br />

0 5<br />

L 4<br />

, 2 parameters<br />

1<br />

0.5<br />

0<br />

−0.5<br />

0 5 10<br />

L 4<br />

, 3 parameters<br />

1<br />

0.5<br />

0<br />

−0.5<br />

0 5 10 15<br />

L 4<br />

, 4 parameters<br />

1<br />

0.5<br />

0<br />

−0.5<br />

0 1 2 3<br />

1<br />

0.5<br />

0<br />

−0.5<br />

0 5<br />

1<br />

0.5<br />

0<br />

−0.5<br />

0 5<br />

Figure 6.3: Wavelets used for the detection of the QRS complex in episode 215 of the<br />

MIT-BIH database.<br />

the designed wavelets tend to perform better in this table. Summarizing, for low-order<br />

wavelet filters, the property to have vanishing moments may thus be beneficial in a<br />

morphologically mixed set, however once a required number of vanishing moments has<br />

been reached one can gain more by maximizing sparsity instead of imposing additional<br />

vanishing moments. The required vanishing moments can be incorporated in the parameterization<br />

as discussed in Chapter 5. The designed wavelets are displayed in Figure 6.3.<br />

Once again wavelets strongly resembling the Daubechies 2 wavelets have been found as<br />

an optimum relatively often.<br />

6.1.2 QT time measurement using designed multiwavelets<br />

As explained in Section 2.3 ECG signals can be separated into various segments. The<br />

QRS segment physically corresponds with the depolarization of the ventricles, which<br />

causes the ventricles to contract, and meanwhile atrial repolarization occurs, but due to<br />

the muscle mass of the ventricles the ventricular activity will be dominant. During the


114 CHAPTER 6. BIOMEDICAL APPLICATIONS OF WAVELET DESIGN<br />

0.3<br />

0.2<br />

0.1<br />

0<br />

−0.1<br />

−0.2<br />

100 200 300 400 500 600 700 800 900 1000<br />

Figure 6.4: Example of an ECG beat from the PTB database. The beat is extended to<br />

have a length that is a power of two. In order to cope with the end points of the signal<br />

when calculating the wavelet transform, periodic extension is employed.<br />

following ST segment the ventricular muscle cells hold their action potential resulting<br />

in an iso-electric segment. The T wave corresponds to the ventricular repolarization.<br />

The start of the Q wave until the end of the T wave thus corresponds to the total<br />

ventricular activation (depolarization and repolarization) and gives the duration of the<br />

electrical systole. As previously discussed in [70] the QT interval is relevant for a large<br />

number of medical applications. QT prolongation is considered an indicator for sudden<br />

cardiac death, see [99]. The QT interval is used to calculate the beat-to-beat variability<br />

of repolarization (BVR) which currently is a an important indicator for some cardiac<br />

pathologies, see [111]. The QT interval is also linked to non-cardiac pathologies such as<br />

diabetic autonomic neuropathy, see e.g. [41].<br />

The accurate measurement of the QT time is a challenging task and was the subject<br />

of the PhysioNet/Computers in Cardiology Challenge 2006 [88]. Many recent methods<br />

[75, 57, 4] attempt to exploit the information that is contained in multiple leads since it<br />

shows a different projection of the same phenomenon (see Section 2.3).<br />

Using the orthogonal multiwavelet design approach from Section 5.3 it is possible<br />

to construct a multiwavelet such that wavelet function in the multiwavelet structure<br />

is designed for the QRS complex and the other wavelet function in the multiwavelet<br />

structure is designed for the T wave, aiming at distinguishing them in the wavelet domain.<br />

In Figure 6.5 the design of a multiwavelet for the simultaneous detection of both<br />

QRS complexes and T waves is displayed. In the top left of the figure the stationary<br />

wavelet decomposition with the first wavelet in the multiwavelet structure is shown.<br />

The white dashed rectangle is a time-scale mask, used during the optimization, where<br />

the time component corresponds to the QRS complex. In the lower left subfigure the<br />

absolute values of wavelet coefficients of this decomposition are displayed. In the top<br />

right of Figure 6.5 the stationary wavelet decomposition using the second wavelet in the<br />

multiwavelet structure is displayed. The rectangle is the time-scale mask used during<br />

the design, where the time component corresponds to the location of the T wave.<br />

Because a large number of different morphologies can manifest in an ECG signal<br />

(see Section 2.3) it is interesting to see whether a multiwavelet can be designed that is


6.1. APPLICATIONS OF WAVELET DESIGN IN CARDIOLOGY 115<br />

Wavelet coefficients 1 Wavelet coefficients 2<br />

abs(Wavelet coefficients 1) abs(Wavelet coefficients 2)<br />

Figure 6.5: Design of a multiwavelet for the simultaneous detection of the QRS complex<br />

and the T wave for the ECG beat in Figure 6.4. In the upper two figures the detail<br />

coefficients are given at full resolution (SWT) from coarse to fine, and in the lower two<br />

figures the corresponding absolute values. The solid rectangle is a time-scale mask that<br />

indicated the region of interest with respect to the QRS complex. The dashed rectangle<br />

is the mask corresponding to the T wave.<br />

applicable in a number of different situations (morphologies). To this end the criterion<br />

for wavelet design has to be modified a bit. Instead of calculating the multiwavelet<br />

decomposition of a single prototype signal and calculating the l 1 or the l 4 norm of this<br />

decomposition, the l 1 or l 4 norms of the decompositions of a vector of prototype signals<br />

is calculated yielding a vector of criterion values (v 1 , v 2 , . . . , v k ), where k is the number of<br />

prototypes. Next these norms have to be merged into a single design criterion. This can<br />

be accomplished by e.g. taking the sums of the entries in this vector ∑ k<br />

l=1 v k, taking the<br />

maximum over the entries of this vector max(v) or taking the product of the sum and<br />

the maximum of the entries of this vector max(v) · ∑k<br />

l=1 v k. In Figure 6.6 a multiwavelet<br />

that has been designed in this way for a vector of prototypes is displayed. A test signal<br />

was constructed for this figure that stitches a number of different morphologies together.<br />

As a reference the decomposition using the Daubechies 2 wavelet was used. Although<br />

a large number of T waves exhibit improved visibility some of them are still not clearly<br />

visible. This applies in particular to the third T wave in Figure 6.6 which has a low<br />

amplitude.<br />

Since the aim is to detect the Q onset and T end points one could argue that the<br />

wavelet is to be designed to detect these points. For this purpose a new multiwavelet


116 CHAPTER 6. BIOMEDICAL APPLICATIONS OF WAVELET DESIGN<br />

D7<br />

D6<br />

D5<br />

D4<br />

D3<br />

D2<br />

D1<br />

D7<br />

D6<br />

D5<br />

D4<br />

D3<br />

D2<br />

D1<br />

D7<br />

D6<br />

D5<br />

D4<br />

D3<br />

D2<br />

D1<br />

(a)<br />

1000 2000 3000 4000 5000 6000 7000 8000<br />

(b)<br />

1000 2000 3000 4000 5000 6000 7000 8000<br />

(c)<br />

1000 2000 3000 4000 5000 6000 7000 8000<br />

(d)<br />

4<br />

2<br />

0<br />

−2<br />

2<br />

0<br />

−2<br />

0.04<br />

0.02<br />

0<br />

−0.02<br />

1.5<br />

1<br />

0.5<br />

1000 2000 3000 4000 5000 6000 7000 8000<br />

Figure 6.6: Design of a multiwavelet for QRS and T wave detection in a range of morphologies.<br />

(a) decomposition with Daubechies 2 wavelet, (b) decomposition with first<br />

multiwavelet, (c) decomposition with second multiwavelet, (d) test signal<br />

was designed, for adapted location of the two masks. This is illustrated in Figure 6.7.<br />

As can be observed in this figure, the detection of the location of the T end point from<br />

this decomposition is not convenient.<br />

For this reason a testing algorithm was developed that performs a multiwavelet transform<br />

on a band-pass filtered ECG signal, considering only a single lead, and then detects<br />

the presence of the QRS complex and the T wave from multiple scales of the stationary<br />

multiwavelet decomposition. The exact location of the QRS onset and T end points is<br />

then determined by a rule-based system that involves the derivative of the single lead<br />

ECG signal. Preliminary results of the current algorithm show that it can detect the location<br />

of the Q onset of 548 beats from the PTB database [16] with a standard deviation<br />

of the error of σ = 15.22ms, for the Q onset point, σ = 30.29ms for the T end point and<br />

σ = 29.87ms for the QT time, with respect to the used gold standard [28] which consists<br />

of manual annotations for beats in the PTB database as published in [25]. It is also<br />

found that the means of the errors are quite high (−2.44ms, −22.82ms and −20.37ms)<br />

respectively and indicate a systematic error that can be corrected at the end of the algorithm<br />

and as a result is not a good measure of the performance of the algorithm. The


6.1. APPLICATIONS OF WAVELET DESIGN IN CARDIOLOGY 117<br />

D7<br />

D6<br />

D5<br />

D4<br />

D3<br />

D2<br />

D1<br />

D7<br />

D6<br />

D5<br />

D4<br />

D3<br />

D2<br />

D1<br />

D7<br />

D6<br />

D5<br />

D4<br />

D3<br />

D2<br />

D1<br />

(a)<br />

1000 2000 3000 4000 5000 6000 7000 8000<br />

(b)<br />

1000 2000 3000 4000 5000 6000 7000 8000<br />

(c)<br />

1000 2000 3000 4000 5000 6000 7000 8000<br />

(d)<br />

1.5<br />

1<br />

0.5<br />

1000 2000 3000 4000 5000 6000 7000 8000<br />

Figure 6.7: Design of a multiwavelet for Q onset and T end detection in a range of<br />

morphologies. (a) decomposition with Daubechies 2 wavelet, (b) decomposition with<br />

first multiwavelet, (c) decomposition with second multiwavelet, (d) test signal<br />

obtained result is not spectacular but most reasonable (see e.g. [75, 57, 4]). Note that<br />

when considering a single lead, the choice of the lead may affect the results [82].


118 CHAPTER 6. BIOMEDICAL APPLICATIONS OF WAVELET DESIGN<br />

6.2 Bias field removal from magnetic resonance images using wavelet<br />

design<br />

The wavelet design procedures from Chapter 5 can also be used for 2D applications, such<br />

as image processing, by taking the Cartesian product of 1D wavelet transforms as in [79].<br />

As a concrete example, bias field removal from magnetic resonance images is considered<br />

as published in [64].<br />

Magnetic Resonance Imaging (MRI), or Nuclear Magnetic Resonance Imaging (NMRI)<br />

as it was called in the early days, is widely used in the medical practice. It provides a<br />

much better contrast for soft tissue than for example x-ray / Computed Tomography<br />

(CT). The technique uses a magnetic field that is generated by coils to align the magnetization<br />

of hydrogen atoms. The alignment of this magnetization is altered with radio<br />

waves such that a rotating magnetic field is created that is picked up by the scanner.<br />

The technology is still undergoing major improvements, both when it comes to hardware<br />

performance (such as field strength increase to improve the visibility of tissue<br />

anomalies) and software development (e.g. for sophisticated MR image processing).<br />

An example of such a new development is 4D acquisition: a time-lapsed 3D image of a<br />

subject.<br />

6.2.1 Radio Frequency inhomogeneities in magnetic resonance images<br />

An important type of artifact of interest is the Radio Frequency (RF) inhomogeneity<br />

or bias field artifact; see e.g. [7]. This type of artifact affects MR images and may<br />

seriously hamper reliable diagnosis in practice. The artifact is caused by the fact that<br />

the intensity of the machine magnetic field varies, depending on the scan sequence, tissues<br />

being imaged and the type of coil being used. Regarding the coil type, mainly surface<br />

coil images suffer from this type of artifact (whereas body coil images suffer less from<br />

this type of artifact, but contain more noise). The effect of RF homogeneity artifacts<br />

is that the image suffers from a non-uniform illumination. This expresses itself as, for<br />

example, extended luminescence or an image shade variation, but also as white stripes<br />

as illustrated in for example Figure 6.8.<br />

6.2.2 Bias field removal in magnetic resonance images<br />

One possible approach to remove this artifact, is to first image a phantom (a container<br />

with a known substance) and then to use this to determine the specific magnetic field<br />

intensity variations, see [32, 113]. From the phantom image a degradation model is<br />

computed, which is then used to account for a bias field in other MRI images obtained<br />

under comparable circumstances. In [64], which is used as a basis for this section, an<br />

approach was described that is intended to facilitate bias field removal from a corrupted<br />

image without prior knowledge of the specific field variations. A closely related approach<br />

with similar goals has previously been described in [7, 8] using classical Butterworth<br />

filters.


6.2. BIAS FIELD REMOVAL FROM MR IMAGES USING WAVELET DESIGN 119<br />

Figure 6.8: Knee MRI corrupted with bias field<br />

6.2.3 Wavelet design for RF inhomogeneity detection<br />

In the literature the RF inhomogeneity is commonly viewed and treated as multiplicative<br />

noise. As discussed in [7, 45] one can apply homomorphic filtering by taking the logarithm<br />

of each pixel (ln(·)) of the corrupted image (C ′ = ln(C)) first, to make the bias field<br />

additive such that it can more easily be removed. In our case the image values are in the<br />

range 0−255, and since the logarithm of zero is undefined the image values are increased<br />

by 1.<br />

In order to suppress the bias field from MR images in a wavelet-based approach, in<br />

[64] wavelets are designed by extending the approach discussed in [69] from 1D to 2D.<br />

This extension is carried out by taking the Cartesian product of 1D wavelet transforms<br />

as in [79]. The design is performed by computing the 2D wavelet decomposition of a<br />

prototype image and optimizing it according to an application dependent place-scale<br />

mask. Since the bias field in the present application covers the whole image, and it is<br />

a low frequency artifact, a scale mask is used which uniformly takes only the coarsest<br />

scales into account.<br />

As a prototype, the natural logarithm of an artificial bias field F B from BrainWeb<br />

[18, 72, 29] was used: F C = ln(F B ). To optimize the wavelet design criterion, the<br />

2D wavelet transform of F C was calculated over 5 scales in the following way. First,<br />

the 1D wavelet transform was calculated in the horizontal direction of each row of F C<br />

giving arrays a1 and d1. Next the 1D wavelet transforms were calculated in the vertical<br />

directions of a1 and d1, giving arrays aa1 and ad1 for a1, and da1 and dd1 for d1.<br />

This splits the image into an approximation part aa1 and a horizontal ad1, a vertical<br />

da1 and a diagonal dd1 detail part. The procedure is then reapplied iteratively on the


120 CHAPTER 6. BIOMEDICAL APPLICATIONS OF WAVELET DESIGN<br />

aa 1 da 1<br />

Image WT on rows a 1 d 1<br />

WT on cols<br />

ad 1 dd 1<br />

WT on rows<br />

Continue 2D wavelet recursion<br />

aa 2 da 2<br />

da 1<br />

WT on cols<br />

a 2<br />

d 2<br />

da 1<br />

ad 2 dd 2<br />

ad 1 dd 1<br />

ad 1 dd 1<br />

Figure 6.9: 2D wavelet transform<br />

approximation coefficient array aa1 only, yielding the arrays aa2, ad2, da2 and dd2. This<br />

procedure is then applied recursively to aa2. This process is illustrated in Figure 6.9.<br />

The depth of this recursion is limited by the size of the image and the length of the filters.<br />

Once the array aa5 is obtained, the l 4 norm of its entries is calculated, providing the<br />

design criterion V from [64] that is to be maximized by tuning the free filter coefficients:<br />

⎛<br />

V (F C ) = max ⎝ ∑ i<br />

∑<br />

j<br />

|aa5 i,j | 4 ⎞<br />

⎠ (6.1)<br />

This procedure assists in creating a wavelet which pushes the energy of the prototype<br />

bias field into the approximation scale. The key idea is to exploit the fact that the bias<br />

field is a slowly varying artifact. This process is illustrated in Figure 6.10. The reader<br />

should note that the scaling of the prototype bias field is not necessarily the same as the<br />

bias field in the actual image. The actual number of scales in the transformation should<br />

be selected in accordance with the actual resolution of the image. In this manner the<br />

low-pass FIR filter<br />

− 0.0230 + 0.0173z −1 + 0.1533z −2 + 0.1349z −3 + 0.5148z −4 + 0.7898z −5 + 0.0455z −6<br />

− 0.2567z −7 + 0.0165z −8 + 0.0219z −9


6.2. BIAS FIELD REMOVAL FROM MR IMAGES USING WAVELET DESIGN 121<br />

Prototype image<br />

wavelet coefficients<br />

|wavelet coefficients|<br />

wavelet and scaling function<br />

Figure 6.10: Wavelet designed to maximize the l 4 -norm of the approximation coefficients<br />

on scale 5. As a prototype the logarithm of an artificial bias field from BrainWeb was<br />

taken. For the optimization n − 1 = 4 free parameters were used, which provided a FIR<br />

filter of length 2n = 10.<br />

and the associated power complementary high-pass FIR filter<br />

0.0219 − 0.0165z −1 − 0.2567z −2 − 0.0455z −3 + 0.7898z −4 − 0.5148z −5 + 0.1349z −6<br />

− 0.1533z −7 + 0.0173z −8 + 0.0230z −9<br />

were obtained.<br />

6.2.4 Filtering MR images with designed wavelets<br />

In order to filter the actual image, the SWT is applied. This technique that is discussed<br />

in Section 3.3 is extended to two dimensions by means of a Cartesian product. The<br />

advantage of this approach is that artifacts that can appear with the regular wavelet<br />

transform (see e.g. Figure 6.11) are avoided when reconstructing the bias field from the<br />

approximation coefficients only. A discussion on artifacts related to such smoothness<br />

issues can be found in, e.g., [107, Chapter 10].


122 CHAPTER 6. BIOMEDICAL APPLICATIONS OF WAVELET DESIGN<br />

0 1 2<br />

3 4 5<br />

Figure 6.11: Original image and reconstructed images with each time the n finest detail<br />

coefficients omitted where n is the title of the image. Note that artifacts are introduced,<br />

due to a lack of smoothness.<br />

Original image C<br />

ln(C+1)<br />

50<br />

50<br />

100<br />

100<br />

150<br />

150<br />

200<br />

200<br />

250<br />

50 100 150 200 250<br />

250<br />

50 100 150 200 250<br />

SWT Approx. coefs.<br />

Wavelet decomposition of ln(C+1)<br />

50<br />

50<br />

100<br />

100<br />

150<br />

150<br />

200<br />

200<br />

250<br />

50 100 150 200 250<br />

250<br />

50 100 150 200 250<br />

Figure 6.12: In clockwise order: the MRI image of a hip, the logarithm of this image, a<br />

5-scale wavelet decomposition of image, and the approximation coefficients of the SWT<br />

of the image.


6.2. BIAS FIELD REMOVAL FROM MR IMAGES USING WAVELET DESIGN 123<br />

Another technicality which requires attention is that when applying homomorphic<br />

filtering, streak artifacts are created at the boundary between tissue and background.<br />

Assume that there is a mask M that indicates the tissue region. In order to avoid these<br />

artifacts normalized convolution [45] is used. The natural logarithm ln(F ) of the bias field<br />

F in the original image O in Figure 6.12 can be reconstructed from the approximation<br />

coefficients of the stationary wavelet decomposition of ln(O)•M, where the • denotes the<br />

Hadamard product, i.e., the entry-wise product. The mask M is filtered in exactly the<br />

same manner giving w(M), where w(·) indicates that it has been filtered using a wavelet<br />

approach. The actual estimated bias field is then obtained by first applying normalized<br />

convolution followed by taking the exponential of each entry exp(·):<br />

( )<br />

w(ln(O) • M)<br />

F = exp<br />

. (6.2)<br />

w(M)<br />

The obtained bias field is shown in the center of Figure 6.13. The original image at the<br />

left side of Figure 6.13 is divided by this bias field to obtain the restored image at the right<br />

side of Figure 6.13. For the criteria discussed in [64] the image needs to be segmented.<br />

For each segment i the mean of the pixel values in the segment µ i are calculated as well as<br />

the standard deviation σ i . Next as previously discussed in [7] the coefficient of variance<br />

cv i = σi<br />

µ i<br />

of each segment and the coefficient of contrast cc i,j = µi<br />

µ j<br />

between each adjacent<br />

pair of segments is calculated. The measures of performance used here are, as discussed<br />

in [64], the minimum of the absolute values of the coefficients of contrast of the adjacent<br />

segments min(bcc|bcc > 0) = min segements i,j adjacent {abs(ln(cc i,j ))}, and the mean of<br />

the negative logarithm of the coefficients of variance mean(lcv) = mean i {− ln(cv i )}. An<br />

overall criterion can be computed as opc = min(bcc|bcc > 0)mean(lcv). In terms of these<br />

performance measures a score of min(bcc|bcc > 0) = 0.392 was obtained for the original<br />

image and min(bcc|bcc > 0) = 0.588 for the reconstructed image. For the reduction in<br />

variance the original image had a score of mean(lcv) = 0.0695 and the reconstructed<br />

image a score of mean(lcv) = 1.024. The respective products are opc = 0.0272 and<br />

opc = 0.602, which constitutes a major improvement for the reconstructed image.<br />

The current approach appears to work well on images with strong small details. The<br />

results for e.g. brain MR images are currently not as good. If the wavelet transform<br />

for these images is calculated on more scales one can see an improvement in the results.<br />

However the maximum scale that can be calculated is limited by both the image size<br />

and the filter size. The relevant components of the image obviously live at the same<br />

coarse scale as the bias field. A possible way to overcome this problem is to design<br />

multiwavelets (Section 5.3) and to simultaneously optimize one wavelet for the bias field<br />

and another orthogonal wavelet for the relevant image such that they become separable.<br />

This approach is more flexible than the current one.


124 CHAPTER 6. BIOMEDICAL APPLICATIONS OF WAVELET DESIGN<br />

Original Bias Restored Image<br />

Figure 6.13: From left to right: the original image, the bias field as detected by the<br />

method and the filtered image after removal of the multiplicative bias field.


Chapter 7<br />

Conclusions and directions for further<br />

research<br />

Technological advancements in the field of biomedical engineering help to lengthen the<br />

lifespan of patients that suffer from cardiovascular related illnesses. One of such advancements<br />

are implantable devices such as pacemakers. Battery life is a critical issue<br />

for bio-implantables. To reduce the power demand the frequency of therapies need to<br />

be reduced, i.e., good sense amplifiers are crucial and the power consumption of these<br />

sense amplifiers needs to be as low as possible. In recent years the technique of wavelet<br />

transforms that offer simultaneous time-frequency resolution with a zoom-in property,<br />

that offer a choice of basis and that can be used to measure the regularity of signals, has<br />

gained popularity in the field of biomedical engineering. Wavelets are a signal processing<br />

technique that can make sense amplifiers in pacemakers more powerful. However,<br />

A/D conversion is power consuming and for low-power applications a low resolution A/D<br />

converter is required. This implies that one should perform as many computations as<br />

possible in the analog domain. It is possible to approximate the wavelet transform by<br />

matching the impulse response of a linear system with the time-reversed, time-shifted<br />

wavelet function. Previously this matching was established by means of Padé approximation.<br />

It is discussed however that this technique has a number of drawbacks:<br />

• The matching is performed in the Laplace domain in the neighborhood of a point<br />

s 0 which has to be chosen in advance.<br />

• Stability of the filter is not automatically guaranteed. The choice s 0 = 0 will help<br />

to obtain stability but unfortunately it is likely to give a poor fit at the beginning<br />

of the impulse response in the region where the wavelet “lives”.<br />

• The choice of the degrees of the numerator and denominator polynomials which<br />

constitute the Padé approximation is not a trivial problem and may strongly influence<br />

the quality of the results.<br />

125


126 CHAPTER 7. CONCLUSIONS AND DIRECTIONS FOR FURTHER RESEARCH<br />

A number of variations to Padé approximation are discussed that attempt to cope with<br />

these problems, but the drawback remains that the quality of the approximation is not<br />

measured directly in the time domain, but in the Laplace domain, with the consequence<br />

that it does not allow for a direct interpretation in system theoretic terms.<br />

In this thesis a novel approach is introduced based on L 2 approximation. This approach<br />

offers a number of conceptual and practical advantages:<br />

• The wavelet transform involves the L 2 -inner product between the wavelet function<br />

and the signal with an arbitrary time shift. The L 2 approximation approach treats<br />

all time instances equal and is appropriate for reliably computing L 2 -inner products.<br />

It allows one to determine an error bound for the approximations obtained in this<br />

way. Optionally it allows for the use of weighing.<br />

• Due to Parsevals’s equality the L 2 -norm allows for a description in both the time<br />

domain and in the Laplace domain, which we use to our advantage.<br />

In order to approximate wavelet functions with the impulse response of linear systems<br />

a parameterization of these impulse responses was proposed that is appropriate for the<br />

approximation of wavelet functions. Additionally it is shown that it is required to enforce<br />

a first vanishing moment in order to avoid a bias in the approximate wavelet transform.<br />

It is shown how the condition for this vanishing moment translates in terms of the<br />

parameters that are used to build linear systems which effectively implement a single<br />

scale of the wavelet transform.<br />

Since the optimization surface corresponding to the wavelet approximation error may<br />

contain multiple local optima, a procedure is proposed to deterministically find a good<br />

starting point for the iterative local search optimization routine. The L 2 approximation<br />

methodology was demonstrated on the Gaussian wavelet, the Morlet wavelet, the Mexican<br />

hat wavelet, the Daubechies 3 wavelet, the Daubechies 7 wavelet and the Coiflet 5<br />

wavelet, of which only one has been successfully implemented directly using the Padé<br />

approach. The results of these approximations are quite satisfactory as a specified approximation<br />

quality is now obtained with approximations of lower order than previously<br />

found with Paé approximation. However it may still happen that the approximation corresponds<br />

to an unsatisfactory local optimum. One may approach this problem in several<br />

ways:<br />

• To develop a different model order reduction technique that specifically attempts<br />

to preserve the shape of the impulse response, rather than to optimize one of the<br />

currently available criteria for model order reduction.<br />

• To employ or develop a global optimization technique for the problem at hand.<br />

General global optimization techniques, such as simulated annealing, may have<br />

the drawback of being computationally intensive. For L 2 approximation, however,<br />

theoretical results are available in discrete-time which guarantee the existence of<br />

only a finite number of local optima.


127<br />

• To develop a practical approach which employs a large number of different starting<br />

points which are well distributed over the class of candidate approximations.<br />

The choice of a specific wavelet may have substantial impact on the performance of<br />

signal processing algorithms such as detection an compression algorithms. This raises<br />

the research question of how to select an appropriate wavelet for the signals and problem<br />

at hand. In order to support this decision it was argued that for applications such as<br />

detection and compression, a sparse representation in the wavelet domain is beneficial.<br />

To obtain sparsity we have persued the principle of maximization of the variance of<br />

the wavelet coefficients. For orthogonal wavelets from filter banks using finite-length<br />

signals, the energy of the detail and approximation coefficients, computed with periodic<br />

extension, will be a constant equalling the energy of the signal. As a result it is possible<br />

to either maximize the variance of the absolute values of the wavelet coefficients which<br />

is shown to boil down to minimizing their l 1 -norm, or to maximize the variance of the<br />

squared wavelet coefficients which comes down to maximization of their l 4 -norm. These<br />

criteria can be used as optimization criteria for the design of wavelets.<br />

As a parameterization of orthogonal wavelets, the existing lattice structure of wavelet<br />

filters in polyphase form was used. Additional constraints to enforce vanishing moments<br />

are worked out and made explicit in terms of the parameterization used. The first vanishing<br />

moment can be built-in by eliminating a parameter. A second vanishing moment<br />

is given in terms of a constraint on the parameters. This novel combination of the<br />

parameterization, the newly introduced design goals and the enforcement of vanishing<br />

moments allows for the design of orthogonal wavelets. It was shown that there exist<br />

local optima on the optimization surface and appropriate measures need to be taken to<br />

arrive at satisfactory results. The applicability of the approach was demonstrated two<br />

examples in Chapter 6: on the detection of the QRS complex in ECG signals and on<br />

a 2D application: bias field removal in MR images. For low-order wavelet filters, the<br />

property to have many vanishing moments can be beneficial in a morphologically mixed<br />

set of signals. However this does not hold in general and once a limited number of vanishing<br />

moments has been reached one can gain more by maximizing sparsity instead of<br />

imposing additional vanishing moments.<br />

One of the major achievements of this thesis is that we have developed a complete<br />

procedure to go from a prototype signal, related to a specific application, to a linear filter<br />

which performs approximate wavelet analysis in the analog domain, ready for implementation<br />

in low-power hardware:<br />

• Wavelet filter design by optimization with a sparsity criterion over a parameterized<br />

class<br />

• Computation of the wavelet and scaling function<br />

• L 2 approximation to construct a linear filter<br />

When one wants to distinguish between morphologies in a given signal, multiwavelets<br />

can be employed. Orthogonal multiwavelets can, as previously discussed in the literature


128 CHAPTER 7. CONCLUSIONS AND DIRECTIONS FOR FURTHER RESEARCH<br />

and further explained in this work, be formulated in polyphase as lossless systems. In<br />

this work a method is introduced that allows for the recursive construction of orthogonal<br />

multiwavelets by employing balanced realizations of lossless systems and the reversed<br />

tangential Schur algorithm. At the heart of the tangential Schur algorithm are linear<br />

fractional transforms that build a a lossless polyphase matrix in state-space form by<br />

means of a recursive procedure. This provides us with a general framework to deal with<br />

orthogonal multiwavelets which encompasses all the existing parameterizations in the<br />

literature, and which has a number of advantages from a computational point of view.<br />

Numerical experience shows shows that a well behaved implementation is feasible, as<br />

orthogonal matrices are employed.<br />

In this work it is also presented how balanced vanishing moments can be built in<br />

to obtain a valid multiwavelet structure that can be used on arbitrary signals without<br />

the need of prefiltering. The first balanced vanishing moment corresponds to an interpolation<br />

condition on the unit circle. Starting from a low-order system with a built-in<br />

balanced vanishing moment, it is shown how the tangential Schur algorithm can be used<br />

to recursively build a lossless system of the desired order, which then gives rise to a multiwavelet<br />

structure. Effectively these ingredients provide a parameterization of orthogonal<br />

multiwavelets with compact support and built-in balanced vanishing moments for which<br />

no parameterization was previously available in the literature. This facilitates the design<br />

of orthogonal multiwavelets. Additionally, a design procedure is developed which<br />

involves masks to highlight different morphologies in a signal to which multiwavelets are<br />

adapted.The potential of this approach is demonstrated on the detection of the Q onset<br />

and T end points in ECG signals for the purpose of QT interval estimation. For image<br />

processing the multiwavelets have been found to be currently too irregular. Additional<br />

balanced vanishing moments are known to enhance the regularity of multiwavelets. In<br />

the future additional interpolation conditions on the unit circle might be imposed in<br />

order to enforce such additional balanced vanishing moments.


Bibliography<br />

[1] Nurettin Acır. Classification of ECG beats by using a fast least square support vector<br />

machines with a dynamic programming feature selection algorithm. Neural Computing &<br />

Applications, 14(4):299–309, December 2005. [cited at p. 111]<br />

[2] P.S. Addison. Wavelet transforms and the ECG: A review. Physiological Measurement,<br />

26:R155–R199, 2005. [cited at p. 25, 27, 64]<br />

[3] P.S. Addison, J.N. Watson, G.R. Clegg, M. Holzer, and C.E. Robertson F. Sterz. Evaluating<br />

arrythmias in ECG signals using wavelet transforms. IEEE Eng. Med. Biol. Mag.,<br />

19(5):104–109, September-October 2000. [cited at p. 64]<br />

[4] R. Almeida, J.P. Martínez, A.P. Rocha, S. Olmos, and P. Laguna. Automatic multilead<br />

VCG based approach for QT interval measurement. In Computers in Cardiology, 2006.<br />

[cited at p. 114, 117]<br />

[5] Rodrigo Varejão Andreão and Jérôme Boudy. Combining wavelet transform and hidden<br />

markov models for ECG segmentation. EURASIP Journal on Advances in Signal Processing,<br />

2007. [cited at p. 27, 45]<br />

[6] I. Androulakis, C. Maranas, and C. Floudas. ffBB: a global optimization method for<br />

general constrained nonconvex problems. Journal of Global Optimization, 7:337–363, 1995.<br />

[cited at p. 86]<br />

[7] E. Ardizzone, R. Pirrone, and O. Gambino. Exponential entropy driven HUM on knee MR<br />

images. In IEEE Eng. Med Biol. Conf., pages 1769 – 1772, September 2005. [cited at p. 118,<br />

119, 123]<br />

[8] E. Ardizzone, R. Pirrone, and O. Gambino. Morphological exponential entropy driven-<br />

HUM. In IEEE Eng. Med. Biol. Conf., pages 3771 – 3774, August 2006. [cited at p. 118]<br />

[9] Jane Austen. Pride and Prejudice. T. Egerton, Whitehall, January 1813. [cited at p. 3]<br />

[10] G.A. Baker Jr. Essentials of Padé Approximants. Academic Press, 1975. [cited at p. 45, 48]<br />

[11] B. Beliczynski, I. Kale, and G.D. Cain. Approximation of FIR by IIR digital filters: an<br />

algorithm based on balanced model reduction. IEEE Trans. Signal Processing, 40(3):532–<br />

542, March 1992. [cited at p. 61]<br />

129


130 BIBLIOGRAPHY<br />

[12] Aharon Ben-Tal and Arkadi Memirovski. Lectures on Modern Convex Optimization.<br />

MPS/SIAM Series on Optimization. MPS-SIAM, 2001. [cited at p. 48]<br />

[13] David H. Bennett. Cardiac Arrhythmias: Practical Notes on Interpretation and Treatment.<br />

Hodder Arnold, 7th edition edition, 2006. [cited at p. 14]<br />

[14] Donald M. Bers. Cardiac excitation-contraction coupling. Nature, 415:198–205, January<br />

2002. [cited at p. 8]<br />

[15] S. Bochner and K. Chandrasekharan. Fourier transforms. Princeton University press,<br />

1949. [cited at p. 15, 17]<br />

[16] R. Bousseljot, D. Kreiseler, and A. Schnabel. Nutzung der EKG-signaldatenbank CAR-<br />

DIODAT der PTB über das internet. Biomedizinische Technik, 40:317, 1995. [cited at p. 14,<br />

116]<br />

[17] R. Bracewell. The Fourier Transform and Its Applications. McGraw-Hill, New York, 3rd<br />

edition edition, 1999. [cited at p. 15]<br />

[18] BrainWeb. Simulated brain database. http://www.bic.mni.mcgill.ca/brainweb/.<br />

[cited at p. 119]<br />

[19] A. Bultheel and M. van Barel. Padé techniques for model reduction in linear system<br />

theory: a survey. Journal of Computational and Applied Mathemathics, 14(3):401–438,<br />

March 1986. [cited at p. 45, 50]<br />

[20] M.J. Burke and M. Nasor. Wavelet based analysis and characterization of the ECG<br />

signal. Journal of Medical Engineering and Technology, 28(2):47–55, March-April 2004.<br />

[cited at p. 27, 64]<br />

[21] E. J. Candès and M. Wakin. An introduction to compressive sampling. IEEE Signal<br />

Processing Magazine, 25(2):21–30, March 2008. [cited at p. 73]<br />

[22] E. J. Candès, M. Wakin, and S. Boyd. Enhancing sparsity by reweighted l1 minimization.<br />

J. Fourier Anal. Appl., 14:877–905, 2007. [cited at p. 73]<br />

[23] Kuei-Fang Chang, Sue-Jen Shih, and Chiou-Mei Chang. Regularity and vanishing moments<br />

of multiwavelets. Taiwanese Journal of Mathematics, 1(3):303–314, September<br />

1997. [cited at p. 91]<br />

[24] J.O. Chapa and R.M. Rao. Algorithms for designing wavelets to match a specified signal.<br />

IEEE Trans. Signal Processing, 48(12):3395– 3406, December 2000. [cited at p. 6, 71]<br />

[25] Ivaylo Christov, Ivan Dotsinsky, Iana Simova, Rada Prokopova, Elina Trendafilova, and<br />

Stefan Naydenov. Dataset of manually measured QT intervals in the electrocardiogram.<br />

Biomedical Engineering Online, 5(31), May 2006. [cited at p. 14, 116]<br />

[26] Charles K. Chui. Wavelets: A Mathematical Tool for Signal Analysis. SIAM monographs<br />

on mathematical modeling and computation. Society for Industrial and Applied Mathematics<br />

(SIAM), Philadelphia, 1997. [cited at p. 26]<br />

[27] C.K. Chui and Q.T. Jiang. Balanced multiwavelets in R s . Mathematics of Computation,<br />

74:1323–1344, 2005. [cited at p. 91, 101, 108]


BIBLIOGRAPHY 131<br />

[28] Jurgen A. H. R. Claassen. The gold standard: not a golden standard. BJM, 330:1121,<br />

May 2005. [cited at p. 116]<br />

[29] D.L. Collins, A.P. Zijdenbos, V. Kollokian, J.G. Sled, N.J. Kabani, C.J. Holmes, and A.C.<br />

Evans. Design and construction of a realistic digital brain phantom. IEEE Trans. Med.<br />

Imaging, 17(3):463–468, 1998. [cited at p. 119]<br />

[30] I. Daubechies. Orthonormal bases of compactly supported wavelets. Comm. on Pure and<br />

Appl. Math., 41(7):909–996, 1988. [cited at p. 25, 30]<br />

[31] Ingrid Daubechies. Ten Lectures on Wavelets. SIAM, 1992. [cited at p. 29, 30]<br />

[32] B.M. Dawant, A.P. Zijdenbos, and R.A. Margolin. Correction of intensity variations in MR<br />

images for computer-aided tissue classification. IEEE Trans. Med. Imaging, 12:770–781,<br />

1993. [cited at p. 118]<br />

[33] David L. Donoho. For most underdetermined systems of linear equations, the minimal<br />

l1-norm near-solution approximates the sparsest near-solution. Comm. Pure Appl. Math.,<br />

59(6):797–829, June 2006. [cited at p. 73]<br />

[34] David L. Donoho and Michael Elad. Optimally sparse representation from overcomplete<br />

dictionaries via l1 norm minimization. Proc. Natl. Acad. Sci., 100(5):2197–2002, March<br />

2003. [cited at p. 73]<br />

[35] Barbara J. Drew. Pitfalls and artifacts in electrocardiography. Cardiology Clinics,<br />

24(3):309–315, August 2006. [cited at p. 10]<br />

[36] W. Einthoven. The different forms of the human electrocardiogram and their signification.<br />

Lancet, 1:853–861, 1912. [cited at p. 10]<br />

[37] Mourad N. El-Gamal and Gordon W. Roberts. A 1.2-V N-P-N-only integrator for logdomain<br />

filtering. IEEE Trans. Circuits Syst. II, 49(4):257–265, April 2002. [cited at p. 43,<br />

44]<br />

[38] D. Esteban and C. Galand. Application of quadrature mirror filters to split-band voice<br />

coding. In IEEE Int. Conf. on Acoustics, Speech and Signal Processing, volume 2, pages<br />

191–195, April 1977. [cited at p. 28, 29, 30]<br />

[39] Richard Feynman. The Feynman Lectures on Physics. Addison Wesley Longman, 1970.<br />

[cited at p. 16]<br />

[40] D. Gabor. Theory of communication. Journal I.E.E., 93(26):429–457, November 1946.<br />

[cited at p. 18]<br />

[41] W.H. Gispen, B. Bravenboer, P.-H. Hendriksen, P.L. Oey, A.C. van Huffelen, and D.W.<br />

Erkelens. Is the corrected QT interval a reliable indicator of the severity of diabetic<br />

autonomic neuropathy? American Diabetes Care, 16(9):1249–1253, 1993. [cited at p. 114]<br />

[42] David E. Goldberg. Genetic Algorithms in Search, Optimization and Machine Learning.<br />

Kluwer Academic Publishers, 1989. [cited at p. 86]<br />

[43] A.L. Goldberger, L.A.N. Amaral, L. Glass, J.M. Hausdorff, P.Ch. Ivanov, R.G. Mark,<br />

J.E. Mietus, G.B. Moody, C.-K. Peng, and H.E. Stanley. Physiobank, physiotoolkit,<br />

and physionet: Components of a new research resource for complex physiologic signals.<br />

Circulation, 101(23):e215–e220, June 2000. [cited at p. 14, 110, 111]


132 BIBLIOGRAPHY<br />

[44] A. Grossman and J. Morlet. Decompostion of hardy functions into square integrable<br />

wavelets of constant shape. Siam J. on Math. Analysis, 15(4):723–736, 1984. [cited at p. 25,<br />

26]<br />

[45] R. Guillemaud. Uniformity correction with homomorphic filtering on region of interest.<br />

In IEEE Int. conf. Image Processing, volume 2, pages 872–875, 1998. [cited at p. 119, 123]<br />

[46] Anubha Gupta, Shiv Dutt Joshi, and Surendra Prasad. A new method of estimating<br />

wavelet with desired features from a given signal. Signal Processing, 85(1):147–161, January<br />

2005. [cited at p. 6, 71]<br />

[47] Sandor A.P. Haddad. Ultra Low-Power Biomedical Signal Processing: An Analog Wavelet<br />

Filter Approach for Pacemakers. PhD thesis, Delt University of Technology, December<br />

2006. [cited at p. 46]<br />

[48] S.A.P. Haddad, S. Bagga, and W.A. Serdijn. Log-domain wavelet bases. In Proceedings of<br />

the 2004 International Symposium on Circuits and Systems, ISCAS ’04, volume 1, pages<br />

1100–1103, May 2004. [cited at p. 4, 43]<br />

[49] S.A.P. Haddad, S. Gieltjes, R.P.M. Houben, and W.A. Serdijn. An ultra low-power dynamic<br />

translinear sense amplifier for pacemakers. In proc. IEEE International Symposium<br />

on Circuits and Systems, 2003. [cited at p. 3, 4, 43]<br />

[50] S.A.P. Haddad, R. Houben, and W.A. Serdijn. Analog wavelet transform employing<br />

dynamic tranlinear circuits for cardiac signal characterization. In Proceedings of the IEEE<br />

International Symposium on Circuits and Systems (ISCAS), volume 1, pages 121–124,<br />

May 2003. [cited at p. 4, 43]<br />

[51] S.A.P. Haddad, R.P.M. Houben, and W.A. Serdijn. The evolution of pacemakers. IEEE<br />

Eng. Med. Biol. Mag., 25(3):38–48, May-June 2006. [cited at p. 3, 7, 8, 10]<br />

[52] S.A.P. Haddad, J.M.H. Karel, R.L.M. Peeters, R.L. Westra, and W.A. Serdijn. Complex<br />

wavelet transform for analog signal processing. In proc.ProRISC’2004, Veldhoven,<br />

November 2004. [cited at p. 64]<br />

[53] S.A.P. Haddad and W.A. Serdijn. Mapping the wavelet transform onto silicon: the dynamic<br />

translinear approach. In Proceedings of the IEEE International Symposium on<br />

Circuits and Systems (ISCAS), volume V, pages 621–624, May 2002. [cited at p. 4, 43]<br />

[54] S.A.P. Haddad, N. Verwaal, R. Houben, and W.A. Serdijn. Optimized dynamic translinear<br />

implementation of the gaussian wavelet transform. In Proceedings of the IEEE International<br />

Symposium on Circuits and Systems, volume I, pages 145–148, May 2004.<br />

[cited at p. 4, 43, 45, 50]<br />

[55] B. Hanzon, M. Olivi, and R.L.M. Peeters. Balanced realizations of discrete-time stable<br />

all-pass systems and the tangential Schur algorithm. Linear Algebra and its Applications,<br />

418(2-3):793–820, October 2006. [cited at p. 6, 72, 89, 90, 91, 92, 93, 94, 98]<br />

[56] Bernard Hanzon and Ralf L.M. Peeters. Balanced parametrizations of stable siso allpass<br />

systems in discrete time. Mathematics of Control, Signals, and Systems (MCSS),<br />

13(3):240–276, September 2000. [cited at p. 91, 92]<br />

[57] D. Hayn, A. Kollmann, and G. Schreier. Automated QT interval measurement from<br />

multilead ECG signals. In Computers in Cardiology, 2006. [cited at p. 114, 117]


BIBLIOGRAPHY 133<br />

[58] W. Heisenberg. Über den anschaulichen inhalt der quantentheoretischen kinematik und<br />

mechanik. Zeitschrift für Physik, 43:172–198, 1927. [cited at p. 18]<br />

[59] J. Willis Hurst. Naming of the waves in the ECG, with a brief account of their genesis.<br />

Circulation, 98:1937–1942, 1998. [cited at p. 12]<br />

[60] M.H. Jager-Geurts, R.J.G. Peters, S.J. van Dis, and M.L. Bots. Hart- en vaatziekten in<br />

Nederland 2006. Technical report, Nederlandse Hartstichting, 2006. [cited at p. 7]<br />

[61] S. Kadambe, R. Murray, and G.F. Boudreaux-Bartels. Wavelet transform-based QRS<br />

complex detector. IEEE Trans. Biomed. Eng., 46(7):838–848, July 1999. [cited at p. 4, 25]<br />

[62] Thomas Kailath. Linear Systems. Prentice Hall Inc., 1980. [cited at p. 18, 20, 22, 53]<br />

[63] A. Kapela, R.D. Berger, A. Achim, and A. Bezerianos. Wavelet variance analysis of highresolution<br />

ECG in patients prone to VT/VF during cardiac electrophysiology studies. In<br />

Proceedings of the 14th International Conference on Digital Signal Processing (DSP 2002),<br />

volume 2, pages 1133–1136, July 2002. [cited at p. 64]<br />

[64] J.M.H. Karel, K. Fischer, R.L. Westra, and R.L.M. Peeters. Measures for the evaluation<br />

of bias field removal procedures in MRI and the design of wavelets for suppression of RF<br />

inhomogeneities. In Proc. Eng.Med.Biol. Conf, 2008. (submitted). [cited at p. 6, 109, 118,<br />

119, 120, 123]<br />

[65] J.M.H. Karel, R.L.M. Peeters, R.L. Westra, S.A.P. Haddad, and W.A. Serdijn. L2-norm<br />

based wavelet approximation for analog implementation. Technical Report M 04-05, Maastricht<br />

University, P.O. Box 616 6200 MD Maastricht The Netherlands, December 2004.<br />

[cited at p. 64]<br />

[66] J.M.H. Karel, R.L.M. Peeters, R.L. Westra, S.A.P. Haddad, and W.A. Serdijn. An L 2-<br />

based approach for wavelet approximation. In Proceedings of the CDC-ECC 2005, 2005.<br />

[cited at p. 55, 57]<br />

[67] J.M.H. Karel, R.L.M. Peeters, R.L. Westra, S.A.P. Haddad, and W.A. Serdijn. Wavelet approximation<br />

for implementation in dynamic translinear circuits. In Proceedings of the 16th<br />

IFAC World Congress. International Federation of Automatic Control, 2005. [cited at p. 53,<br />

55]<br />

[68] J.M.H. Karel, R.L.M. Peeters, R.L. Westra, S.A.P. Haddad, and W.A. Serdijn. Wavelet<br />

design for analog implementation. In 25th Benelux Meeting on Systems and Control,<br />

Heeze, The Netherlands, March 2006. [cited at p. 6]<br />

[69] J.M.H. Karel, R.L.M. Peeters, R.L. Westra, K.M.S. Moermans, S.A.P. Haddad, and W.A.<br />

Serdijn. Optimal discrete wavelet design for cardiac signal processing. In Proc. 27th<br />

Ann. Int. Conf. of the IEEE Enigneering in Medicine and Biology Society (EMBC 2005),<br />

Shanghai, China, September 1-4. IEEE, September 2005. [cited at p. 6, 71, 73, 119]<br />

[70] Joël Karel, Ralf Peeters, Ronald Westra, Sandro Haddad, and Wouter Serdijn. QT interval<br />

measurement in cardiac signal processing with multiwavelets. In First Dutch Conference<br />

on Bio-Medical Engineering, page 184, January 2007. [cited at p. 114]<br />

[71] S. Kirkpatrick, C. D. Gelatt, and M. P. Vecchi. Optimization by simulated annealing.<br />

Science, 220(4598):671–680, 1983. [cited at p. 86]


134 BIBLIOGRAPHY<br />

[72] R.K.-S. Kwan, A.C. Evans, and G.B. Pike. An extensible MRI simulator for postprocessing<br />

evaluation. In Visualization in Biomedical Computing (VBC’96), volume 1131<br />

of Lecture Notes in Computer Science, pages 135–140, 1996. [cited at p. 119]<br />

[73] J.C. Lagarias, J. A. Reeds, M. H. Wright, and P. E. Wright. Convergence properties<br />

of the Nelder-Mead simplex method in low dimensions. SIAM Journal of Optimization,<br />

9(1):112–147, 1998. [cited at p. 86]<br />

[74] Cornelius Lanczos. Applied analysis. Prentice-Hall Mathematics Series. Prentice Hall,<br />

Inc., 1956. [cited at p. 54, 90]<br />

[75] P. Langley, F.E. Smith, S.T. King, D. Zheng, A.J. Haigh, and A. Murray. Fully automated<br />

computer measurement of QT interval from the 12-lead electrocardiogram. In Computers<br />

in Cardiology, 2006. [cited at p. 114, 117]<br />

[76] J. Lebrun and M. Vetterli. Balanced multiwavelets theory and design. IEEE Trans. Signal<br />

Processing, 46(4):1119–1125, April 1998. [cited at p. 6, 40, 72, 89, 91, 108]<br />

[77] J. Lebrun and M. Vetterli. High-order balanced multiwavelets: theory, factorization, and<br />

design. IEEE Trans. Signal Processing, 49(9):1918–1930, 2001. [cited at p. 6, 40, 72, 89, 91,<br />

101, 103, 108]<br />

[78] Cuiwei Li, Chongxun Zheng, and Changfeng Tai. Detection of ECG characteristic<br />

points using wavelet transforms. IEEE Trans. Biomed. Eng., 42(1):21–28, January 1995.<br />

[cited at p. 4, 25]<br />

[79] Stéphane Mallat. A theory for multiresolution signal decomposition: The wavelet representation.<br />

IEEE Trans. Pattern Anal. Machine Intell., 11(7):674 – 693, July 1989.<br />

[cited at p. 25, 31, 101, 118, 119]<br />

[80] Stéphane Mallat. A Wavelet Tour of Signal Processing. Academic Press, 1999.<br />

[cited at p. 18, 26, 27, 32, 33, 34, 51, 72]<br />

[81] Stéphane Mallat. A Wavelet Tour of Signal Processing: The Sparse Way. Academic Press,<br />

2008. [cited at p. 73]<br />

[82] J.P. Martínez, R. Almeida, S. Olmos, A.P. Rocha, and P. Laguna. Stability of QT measurements<br />

in the PTB database depending on the selected lead. In Computers in Cardiology,<br />

2006. [cited at p. 117]<br />

[83] Juan Pablo Martínez, Rute Almeida, Salvador Olmos, Ana Paula Rocha, and Pablo Laguna.<br />

A wavelet-based ECG delineator: Evaluation on standard databases. IEEE Trans.<br />

Biomed. Eng., 51(4):570–581, April 2004. [cited at p. 4, 25]<br />

[84] Hamish D. Meikle. A New Twist to Fourier Transforms. Technology & Industrial Arts.<br />

Wiley-VCH, 2004. [cited at p. 17]<br />

[85] Y. Meyer. Constructions de bases orthonormées d’ondelettes. Rev. Mat. Iberoamericana,<br />

4:31–39, 1988. [cited at p. 25]<br />

[86] Fred Mintzer. Filters for distortion-free two-band multirate filter banks. IEEE Trans.<br />

Acoustics, Speech and Signal Proc., 33(3):626–630, June 1985. [cited at p. 30]


BIBLIOGRAPHY 135<br />

[87] F. Mochimaru, Y. Fujimoto, and Y. Ishikawa. Detecting the fetal electrocardiogram by<br />

wavelet theory-based methods. Progress in Biomedical Research, 7(3):185–193, September<br />

2002. [cited at p. 4, 25]<br />

[88] G.B. Moody, H. Koch, and U. Steinhoff. The physionet/computers in cardiology challenge<br />

2006: QT interval measurement. In Computers in Cardiology, 2006. [cited at p. 14, 114]<br />

[89] B.C. Moore. Principal component analysis in linear systems: Controllability, observability,<br />

and model reduction. IEEE Trans. Automat. Contr., 26(1):17–32, 1981. [cited at p. 60]<br />

[90] Jan Mulder, Wouter Serdijn, Albert C. van der Woerd, and Arthur H.M. van Roermund.<br />

Dynamic translinear circuits - an overview. Analog Integrated Circuits and Signal Processing,<br />

22(2-3):111–126, 2000. [cited at p. 43, 44]<br />

[91] Guy P. Nason and Bernard W. Silverman. The stationary wavelet transform and some<br />

statistical applications. In Anestis Antoniadis and Georges Oppenheim, editors, Lecture<br />

Notes in Statistics: Wavelets and Statistics, volume 103 of Lecture Notes in Statistics,<br />

pages 281–299. Springer-Verlag, 1995. [cited at p. 33, 37, 38, 110]<br />

[92] N. Neretti and N. Intrator. An adaptive approach to wavelet filter design. In Proceedings<br />

of the IEEE international workshop on neural networks for signal processing, September<br />

2002. [cited at p. 6, 25, 71]<br />

[93] Nikolay Nikolaev and Atanas Gotchev. ECG signal denoising using wavelet domain<br />

Wiener filtering. In European Singal Processing Conference (Eusipco), September 2000.<br />

[cited at p. 111]<br />

[94] H. Nyquist. Certain topics in telegraph transmission theory. Proceedings of the IEEE,<br />

90(2):280–305, February 2002. [cited at p. 15]<br />

[95] R.L.M. Peeters, J.M.H. Karel, R.L. Westra, S.A.P. Haddad, and W.A. Serdijn. Multiwavelet<br />

design for cardiac signal processing. In Proc. 28th Ann. Int. Conf. of the IEEE<br />

Enigneering in Medicine and Biology Society (EMBC 2006), New York City, August 30 -<br />

September 3, pages 1682–1685, 2006. [cited at p. 89, 91, 101, 102]<br />

[96] L. Pernebo and L.M. Silverman. Model reduction via balanced state space representations.<br />

IEEE Trans. Automat. Contr., 27(2):382–387, 1982. [cited at p. 60]<br />

[97] Jan Willem Polderman and Jan C. Willems. Introduction to Mathematical Systems Theory,<br />

A Behavioral Approach. Texts in applied mathematics: 26. Springer, 1997. [cited at p. 53]<br />

[98] Ivo Provazník, Jiří Kozumplík, Jana Bardoňová, Marie Nováková, and Zuzana Nováková.<br />

Wavelet transform in ECG signal processing. In Proceedings of the 15th International<br />

Conference BIOSIGNAL 2000, pages 21–25. Vutium Press Brno, 2000. [cited at p. 4, 25]<br />

[99] P.E. Puddu and M.G. Bourassa. Prediction of sudden death from QTc interval prolongation<br />

in patients with chronic ischemic disease. Journal of Electrocardiology, 19:203–212,<br />

1986. [cited at p. 114]<br />

[100] R.M. Rangayyan. Biomedical Signal Analysis: A case-study approach, volume IEEE Press<br />

series on biomedical engineering. IEEE Press, 2002. [cited at p. 10]<br />

[101] Philippe Ravier and Olivier Buttelli. Robust detection of QRS complex using Klauder<br />

wavelets. In EUSIPCO 2004, 2004. [cited at p. 6, 25, 71]


136 BIBLIOGRAPHY<br />

[102] J. Sahambi, S. Tandon, and R. Bhatt. Using wavelet transforms for ECG characterization.<br />

IEEE Eng. Med. Biol. Mag., 16(1):77–83, 1997. [cited at p. 4, 25, 27, 45]<br />

[103] S.C. Saxena, V. Kumar, and S.T. Hamde. QRS detection using new wavelets. Journal of<br />

Medical Engineering and Technology, 26(1):7–15, January-February 2002. [cited at p. 4, 71]<br />

[104] C.E. Shannon. Communication in the presence of noise. Proceedings of the IRE, 37(1):10–<br />

21, January 1949. [cited at p. 15]<br />

[105] D. A. Sherman. An introduction to wavelets with electrocardiology applications.<br />

Herzschrittmachertherapie und Elektrophysiologie, 9:42–52, 1998. [cited at p. 25, 111]<br />

[106] M.J.T. Smith and T. P. Barnwell III. A procedure for designing exact reconstruction filter<br />

banks for tree-structured subband coders. In IEEE Int. Conf on Acoustics, Speech and<br />

Signal Processing, volume 9, pages 421–424, March 1984. [cited at p. 29]<br />

[107] Gilbert Strang and Truong Nguyen. Wavelets and Filter Banks. Wellesley-Cambridge<br />

Press, 1996. [cited at p. 6, 25, 27, 28, 29, 31, 32, 33, 34, 35, 40, 71, 72, 75, 76, 79, 89, 100, 121]<br />

[108] Gilbert Strang and Vasily Strela. Orthogonal multiwavelets with vanishing moments.<br />

Journal of Optical Engineering, 33:2104–2107, 1994. [cited at p. 91]<br />

[109] Vasily Strela. Multiwavelets: Regularity, orthogonality and symmetry via two-scale similarity<br />

transform. Studies in Applied Mathematics, 98:335–354, 1997. [cited at p. 91]<br />

[110] Jacques Theys. Joint Spectral Radius: theory and approximations. PhD thesis, Université<br />

Catholique de Louvain, May 2005. [cited at p. 32]<br />

[111] M.B. Thomsen, S.C. Verduyn, M. Stengl, J.D.M. Beekman, G. de Pater, J. van Opstal,<br />

P.G.A. Volders, and M.A. Vos. Increased short-term variability of repolarization predicts<br />

d-sotalol-induced torsades de pointes in dogs. Circulation, 110:2453–2459, 2004.<br />

[cited at p. 114]<br />

[112] P. E. Tikkanen. Nonlinear wavelet and wavelet packet denoising of electrocardiogram<br />

signal. Biological Cybernetics, 80(4):259 – 267, April 1999. [cited at p. 4, 25]<br />

[113] M. Tincher, C.R. Meyer, R. Gupta, and D.M. Williams. Polynomial modelling and reduction<br />

of RF body coil spatial inhomogeinity in MRI. IEEE Trans. on Med. Imaging,<br />

12:361–365, 1993. [cited at p. 118]<br />

[114] Hüseyin Tirtoma, Mehmet Engin, and Erkan Zeki Engin. Enhancement of time-frequency<br />

properties of ecg for detecting micropotentials by wavelet transform based method. Expert<br />

Systems with Applications, 34(1):746–753, January 2008. [cited at p. 111]<br />

[115] Richard J. Vaccaro. Digital Control: A State-Space Approach. McGraw-Hill Series in<br />

Electrical an Computer Engineering. McGraw-Hill, 1995. [cited at p. 22, 62]<br />

[116] P.P. Vaidyanathan. Theory and design of m-channel maximally decimated quadrature<br />

mirror filters with arbitrary m, having the perfect-reconstruction property. IEEE Trans.<br />

Acoustics, Speech and Signal Proc., 35(4):476–492, April 1987. [cited at p. 41, 72, 75, 89, 100]<br />

[117] P.P. Vaidyanathan and Zinnur Doǧanata. The role of lossless systems in modern digital<br />

signal processing: analog tutorial. IEEE Trans. Education, 32(3):181–197, August 1989.<br />

[cited at p. 6, 29, 30, 41, 42, 72, 89, 100]


BIBLIOGRAPHY 137<br />

[118] C.F. van Loan. Computing integrals involving the matrix exponential. IEEE Trans.<br />

Automat. Contr., 23:395–404, 1978. [cited at p. 62]<br />

[119] Martin Vetterli. Filter banks allowing perfect reconstruction. Signal Processing, 10:219–<br />

244, 1986. [cited at p. 40, 89]<br />

[120] Martin Vetterli and Cormac Herley. Wavelets and filter banks: relationships and new<br />

results. In IEEE Conference on Acoustics, Speech and Signal Processing, volume 3, pages<br />

1723–1726, 1990. [cited at p. 34]<br />

[121] Jan C. Willems and Paolo Rapisarda. Balanced state representation from higher-order<br />

models. In Proceedings Equadiff 2003, Hasselt, Belgium, 2003. [cited at p. 60]<br />

[122] Mark A. Wood and Kenneth A. Ellenbogen. Cardiac pacemakers from the patient’s perspective.<br />

Circulation, 105(18):2136–2138, 2002. [cited at p. 3, 7, 8]<br />

[123] World Health Organization. Cardiovascular diseases - fact sheet 317.<br />

http://www.who.int/mediacentre/factsheets/fs317/en/index.html, February 2007.<br />

[cited at p. 3, 7]<br />

[124] Douglas P. Zipes and José Jalife. Cardiac electrophysiology: from cell to bedside. Saunders,<br />

2000. [cited at p. 9]


Summary<br />

Each year a large number of people decease due to cardiovascular disorders. Due to<br />

medical advancements the contribution of these disorders to the total mortality is reduced.<br />

One of these advancements is a joint medical and engineering development: the<br />

implantable devices, among which the pacemaker. Early pacemakers electro-stimulated<br />

the heart at a fixed rate, ensuring that the heart contracts according to that rate and<br />

pumps blood throughout the body. This approach however has a number of drawbacks.<br />

First of all, the continuous stimulation of the heart is power consuming, which is a major<br />

issue for a device that is surgically inserted into the body and is not easily rechargeable.<br />

Secondly, unnecessary stimulation of the heart can be harmful. Thirdly, the fixed rate of<br />

the pacemaker can interfere with the intrinsic rate of the heart. Nowadays pacemakers<br />

are equipped with a sensing circuit that monitors the electrical currents on the heart.<br />

To accomplish this a monitoring mechanism is required that performs signal processing,<br />

leading to a decision step, in which the decision is made whether or not to stimulate the<br />

heart. Considering that this monitoring circuit is always active, it must be ensured that<br />

this circuit is not power-consuming. Obviously this circuit must also be robust since a<br />

patient’s life depends on it.<br />

A relatively young signal processing technique that is interesting for the use in pacemakers<br />

is the so-called “wavelet transform”. With this technique a signal can simultaneously<br />

be represent in both the time- and frequency-domain. It is easy to implement<br />

in a computer. However, a computer generally operates in the digital domain, whereas<br />

the sensor information is in the analog domain. An analog signal thus has to be converted<br />

to the digital domain, which is a power-consuming operation. An energy saving<br />

solution is to implement the wavelets in the analog domain, and in this manner to reduce<br />

the amount of analog to digital conversion. To achieve this, wavelets have to be<br />

approximated by means of linear systems that can be applied in microelectronics. This<br />

is not a trivial task. In the current work it is discussed why L 2 approximation is a<br />

relevant technique. A complete approach is discussed to approximate wavelet functions<br />

(associated with both continuous and discrete-time wavelets) with this technique. Since<br />

the optimization surface may contain various local optima, it is discussed how a good<br />

starting point for a local search algorithm can be found. Various examples illustrate the<br />

139


140 SUMMARY<br />

flexibility of this approach.<br />

The simultaneous time-frequency representation characteristic is not the only distinction<br />

between wavelets and Fourier transforms. Unlike Fourier transforms, wavelet<br />

transforms offer a choice of bases. Two criteria, that operate in the wavelet domain, are<br />

introduced to determine how good a certain wavelet is for compressing or detecting a<br />

given signal. These criteria are then used, along with a parameterization of orthogonal<br />

wavelets based on “polyphase filters” and the ‘lattice structure”, to design custom<br />

wavelets for an application at hand. Not only scalar wavelets are of interest, but also<br />

multiwavelets. These involve multiple orthogonal wavelet and scaling functions that enable<br />

them to separate orthogonal components in a signal. A parameterization in terms of<br />

“lossless systems” is introduced for these multiwavelets. This parameterization is more<br />

general than the parameterizations that are known from the literature. For a number<br />

of these parameterizations it is discussed how these follow from the introduced parameterization<br />

as special cases. For the new parameterization it is additionally discussed<br />

haw balanced vanishing moments can be built-in, which is required in order to use these<br />

designed multiwavelets directly on measured signals. These balanced vanishing moments<br />

were not explicitly build into earlier parameterizations.<br />

To demonstrate the potential of the discussed techniques, three examples are worked<br />

out. Firstly the designed scalar wavelets are used to detect the QRS complex in an ECG.<br />

Experiments show that the designed wavelets indeed offer advantages. An interesting<br />

observation is that the Daubechies 2 wavelets is often found as an optimum. As a second<br />

application designed multiwavelets are used to simultaneously distinguish the Q and the<br />

T peak in ECGs. A rule-based decision algorithm that has been designed as a proof<br />

of principle readily shows promising results. As a third application wavelet design is<br />

used to facilitate the processing of MR images. Using wavelet filtering, low-frequency<br />

multiplicative noise is successfully removed from images of the pelvis and the knee. The<br />

current technique is not as successful on MR images of the brain. Recommendations to<br />

increase the performance are made.


Samenvatting<br />

Jaarlijks sterft een groot aantal mensen ten gevolge van cardiovasculaire aandoeningen.<br />

Door vooruitgang op het gebied van de medische wetenschap, waaronder meer specifiek<br />

implanteerbare apparaten, wordt het aandeel in de totale sterfte verkleind. Een van de<br />

eerste van dergelijke implanteerbare apparaten is de pacemaker. De vroege uitvoeringen<br />

geven met een vast ritme elektrische pulsen aan het hart, waardoor dit met het vastgestelde<br />

ritme samenknijpt en bloed door het lichaam pompt. Hier kleeft echter een<br />

aantal nadelen aan. Zo kost het voortdurend stimuleren van het hart veel energie, wat<br />

een groot probleem vormt omdat een dergelijk apparaat chirurgisch in het lichaam van<br />

de patiënt is geplaatst en daardoor moeilijk opgeladen kan worden. Ook kan het onnodig<br />

stimuleren van het hart schadelijk zijn. Daarnaast kan het vaste ritme van de pacemaker<br />

interfereren met de frequentie waarop het hart zelf wil samentrekken. Tegenwoordig<br />

zijn pacemakers daarom uitgerust met sensoren die de elektrische stromen op het hart<br />

registreren die vervolgens worden gebruikt in een beslissingsstap waarin wordt besloten<br />

of het hart gestimuleerd moet worden of niet. Dit vereist een monitoringsmechanisme<br />

dat signaalverwerkingstaken uitvoert. Aangezien het monitoring circuit altijd actief is,<br />

is het belangrijk dat dit zuinig met de energie omgaat. Daarnaast moet het de juiste<br />

beslissingen nemen, omdat deze van levensbelang kunnen zijn.<br />

Een relatief recente signaalverwerkingstechniek die interessant is voor het gebruik in<br />

pacemakers is de zogenaamde "wavelet transformatie". Deze techniek maakt het mogelijk<br />

om een signaal tegelijkertijd in tijd en frequentie af te beelden. Het is gemakkelijk om<br />

deze techniek in een computer te implementeren. Echter een computer werkt digitaal en<br />

de sensor informatie in pacemakers is analoog. Er moet dus een analoog signaal omgezet<br />

worden in een digitaal signaal wat veel energie vergt. Een energiezuinige oplossing is om<br />

de wavelets op een analoge manier te implementeren en zo de analoog/digitaal omzetting<br />

te beperken. Om dit te doen moeten de wavelet functies benaderd worden met behulp<br />

van lineaire systemen die in de microelektronica toepasbaar zijn. Dit is echter geen<br />

triviale opgave. In het huidige werk wordt beargumenteerd waarom L 2 approximatie<br />

een goede aanpak is hiervoor. Een complete aanpak om wavelet functies (zowel van<br />

continue als discrete wavelets) met deze techniek te benaderen is uitgewerkt. Aangezien<br />

het optimalisatieoppervlak diverse locale optima kan bevatten, wordt er besproken hoe<br />

141


142 SAMENVATTING<br />

een goed startpunt voor een locale zoektechniek kan worden gevonden. Met diverse<br />

voorbeelden wordt de flexibiliteit van de aanpak aangetoond.<br />

Buiten het feit dat wavelets een signaal tegelijkertijd in termen van tijd als frequentie<br />

laten zien is er nog een ander duidelijk verschil met een klassieke techniek als Fourier<br />

transformaties: er zijn geen vaste basissen en er is keuzevrijheid. Dit geeft meteen een<br />

keuzeprobleem. Er worden twee criteria behandeld die in het wavelet domein bepalen<br />

hoe goed een gegeven wavelet is om een bepaald signaal te comprimeren of te detecteren.<br />

Daarnaast worden deze criteria met een parameterizatie op basis van “polyphase filters”<br />

en de “lattice” structuur van discrete, orthogonale wavelets, gebruikt om wavelets te<br />

ontwerpen. Naast reguliere wavelets zijn ook zogenaamde “multiwavelets” interessant.<br />

Deze beschikken over meerdere orthogonale wavelet en schalingsfuncties en kunnen diverse<br />

orthogonale componenten in een signaal onderscheiden. Voor deze multiwavelets<br />

is een parameterisatie opgezet in termen van zogenaamde “lossless” systemen. Deze parameterisatie<br />

is algemener dan eerdere parameterisaties die bekend zijn uit de literatuur.<br />

Van een aantal van deze bestaande parameterisaties wordt besproken hoe ze volgen uit<br />

de geïntroduceerde, algemenere parameterisatie. Daarnaast wordt besproken hoe met de<br />

geïntroduceerde parameterisatie gebalanceerde momenten kunnen worden ingebouwd,<br />

hetgeen noodzakelijk is om de, met deze parameterisatie, ontworpen multiwavelets direct<br />

op gemeten signalen toe te passen. Deze gebalanceerde momenten werden in eerdere<br />

parameterisaties niet ingebouwd.<br />

Om de kracht van de besproken technieken te demonstreren wordt een drietal voorbeelden<br />

uitgewerkt. Als eerste worden ontworpen wavelets gebruikt om het QRS complex<br />

in een hartsignaal te detecteren. Uit numerieke gegevens blijkt dat deze ontworpen<br />

wavelets inderdaad voordelen bieden. Een opmerkelijk resultaat is dat een bepaalde<br />

wavelet, nl. de Daubechies 2 wavelet dikwijls als optimum gevonden wordt. Als tweede<br />

applicatie worden ontworpen multiwavelets gebruikt om tegelijkertijd de Q piek en de T<br />

piek in een hartsignaal te onderscheiden. Een regelgebaseerd detectiealgoritme dat als<br />

voorbeeld is ontworpen laat reeds bemoedigende resultaten zien. Als derde toepassing<br />

worden wavelets ontworpen en gebruikt voor het verwerken van afbeeldingen afkomstig<br />

van een MRI scanner. Met behulp van deze technieken wordt succesvol laagfrequente<br />

multiplicatieve ruis verwijderd bij afbeeldingen van het bekken en de knie. Op MRI<br />

afbeeldingen van hersenen is de huidige techniek minder succesvol, echter aanbevelingen<br />

om dit verder te verbeteren worden gedaan.


Curriculum Vitae<br />

1978 Born on 10 April 1978 in Maastricht, The Netherlands<br />

1990–1997 Secondary School, Jeanne d’Arc College Maastricht, HAVO/VWO<br />

1997–2001 Master of Science in Knowledge Engineering, business mathematics<br />

major. Joint programme of the Maastricht University (The Netherlands)<br />

and Limburgs Universitair Centrum (Nowadays Hasselt University,<br />

Belgium). Obtained the Belgian diploma “Kandidaat Informatica”<br />

(Computer Science) from the Limburgs Universitair Centrum<br />

in 1999. Attended a special four-week programme on knowledge<br />

engineering and computer science in 1999 at Baylor University,<br />

Waco, Texas. Master thesis at Medtronic Bakken Research Center<br />

b.v., Maastricht, The Netherlands, on the mathematical modeling of<br />

atrial fibrillation, completing a five years dual track at the Maastricht<br />

University in four years in 2001.<br />

2001–2004 Joint position at MaTeUM b.v., Maastricht, The Netherlands as a developer<br />

and at the Maastricht University, Faculty of General Sciences<br />

for research on simulation.<br />

2004–2008 Ph.D. Student on the STW funded BioSens project at the Maastricht<br />

University, Faculty of Humanities and Sciences, MICC, Department<br />

of Mathematics.<br />

2007–2008 Half-time position as a lecturer at the Maastricht University, Faculty<br />

of Humanities and Sciences, MICC, Department of Mathematics.<br />

2008–current Assistant Professor at the Maastricht University, Faculty of Humanities<br />

and Sciences, Department of Knowledge Engineering.<br />

143


List of Symbols<br />

and Abbreviations<br />

Symbol Description Definition<br />

• Hadamard product page 123<br />

∗ convolution operator page 17<br />

†<br />

Hermitian transpose page 42<br />

↓2 downsampling by 2; (↓2x) k = x 2k page 28<br />

A system matrix for a linear system in state-space representation<br />

page 21<br />

a k approximation or scaling coefficients page 31<br />

B input matrix for a linear system in state-space representation<br />

page 21<br />

b k detail or wavelet coefficients page 31<br />

C output matrix for a linear system in state-space representation<br />

page 21<br />

c k scaling filter coefficients page 30<br />

δ[n] Kronecker delta page 18<br />

δ(t) Dirac delta page 18<br />

D direct feed-through matrix for a linear system in statespace<br />

page 21<br />

representation<br />

d k wavelet filter coefficients page 30<br />

e n n th standard basis vector page 99<br />

exp element-wise exponential page 123<br />

F U,V mapping for a proper rational matrix page 92<br />

H(E) Halmos extension page 95<br />

H Hankel matrix page 61<br />

h(t) impulse response page 20<br />

H(s) transfer function page 20<br />

H 0(z) wavelet low-pass filter page 28<br />

H 1(z) wavelet high-pass filter page 28<br />

H p(z) polyphase matrix in z page 35<br />

145


146 LISTS OF SYMBOLS AND ABBREVIATIONS<br />

Symbol Description Definition<br />

i complex number page 16<br />

1 A(x) indicator function page 54<br />

L Laplace transform page 18<br />

ln Element-wise natural logarithm page 119<br />

ω frequency variable page 16<br />

φ phase angle page 16<br />

u(t) input for a system page 19<br />

ϕ(t) scaling function page 31<br />

ϕ(t) multi scaling function page 40<br />

ψ(t) wavelet function page 26<br />

˜ψ(t) time-reversed and time-shifted wavelet function page 47<br />

ˇψ(t) causal time-reversed and time-shifted wavelet function page 47<br />

ψ(t) multi wavelet function page 40<br />

σ scale page 26<br />

T linear fractional transformation page 93<br />

W (τ, σ) wavelet transform page 26<br />

x(t) state vector page 21<br />

y(t) output for a system page 19<br />

Y (iω) Laplace transform of y(t) for s = iω page 18<br />

Y (ω) Fourier transform of y(t) page 18<br />

Y (s) Laplace transform of y(t) page 18<br />

Z Z-transform page 19<br />

Abbreviation Description Definition<br />

3D three-dimensional page 10<br />

A/D analog-to-digital page 15<br />

AF atrial fibrillation page 14<br />

AV atrio-ventricular page 9<br />

BVR beat-to-beat variability of repolarization page 114<br />

CT computed tomography page 118<br />

CWT continuous wavelet transform page 26<br />

DC direct current page 21<br />

DTL dynamic translinear page 43<br />

DWT discrete wavelet transform page 27<br />

ECG electrocardiogram page 10<br />

EKG electrocardiogram page 10<br />

FIR finite impulse response page 23<br />

Hz Hertz page 16<br />

IECG intracardiac electrocardiogram page 10<br />

LA left atrium page 8<br />

LFT linear fractional transformations page 93<br />

LTI linear time-invariant page 19<br />

LV left ventricle page 8


147<br />

Abbreviation Description Definition<br />

MA moving average page 23<br />

MIMO multi-input multi-output page 21<br />

MIT-BIH Massachusetts Institute of Technology - Beth Israel page 14<br />

Deaconess Medical Center<br />

MR magnetic resonance page 118<br />

MRA multi-resolution analysis page 31<br />

MRI magnetic resonance imaging page 118<br />

NMRI nuclear magnetic resonance imaging page 118<br />

pdf probability density function page 27<br />

RA right atrium page 8<br />

RF radio frequency page 118<br />

RV right ventricle page 8<br />

SA sino-atrial page 9<br />

SISO single-input single-output page 20<br />

VF ventricular fibrillation page 14<br />

ZOH zero-order-hold page 18


Index<br />

l 1 -norm, 73<br />

l 4 -norm, 73<br />

A/D converter, 15, 43<br />

action potential, 9<br />

admissability condition, 26<br />

aliasing, 28<br />

all-pass systems, 41, 90<br />

alternating flip, 30<br />

alternating signs, 29<br />

analog, 15<br />

analysis, 16<br />

angular frequency, 16<br />

anisotropy, 10<br />

approximation<br />

L 2 , 51<br />

coefficients, 28<br />

error, 34<br />

Padé, 48<br />

approximation space, 31<br />

arrhythmia, 10<br />

asymptotic stability, 20, 91<br />

atrial fibrillation, 14<br />

atrio-ventricular node, 9<br />

atrium, 8<br />

balance and truncate, 62<br />

balancing, 60, 98, 101, 108<br />

bandlimited, 16<br />

bandwidth, 16<br />

bias field artifact, 118<br />

Bode plot, 16<br />

Bohl functions, 53<br />

canonical form, 21<br />

cardiovascular, 7<br />

cascade of filter banks, 31, 75<br />

Cauchy-Schwarz inequality, 52<br />

causality, 20<br />

closure, 31<br />

coefficient of contrast, 123<br />

coefficient of variance, 123<br />

coefficients<br />

approximation, 28<br />

scaling, 28<br />

wavelet, 28<br />

compact support, 29, 40<br />

Computed Tomography, 118<br />

conjugate transpose, 42<br />

conservation of energy, 29, 72<br />

continuous<br />

Fourier transform, 16<br />

time, 15<br />

wavelet transform, 26<br />

controllability Grammian, 60<br />

controllability matrix, 61<br />

controllable companion form, 61<br />

convolution, 17, 45<br />

convolution kernel, 51<br />

critical sampling, 28, 36<br />

Daubechies wavelets, 34<br />

decision stage, 8<br />

decomposition, 31<br />

defibrillator, 14<br />

depolarization, 8<br />

depolarized, 9<br />

design, wavelet, 72<br />

detail<br />

coefficients, 28<br />

signal, 33<br />

detail space, 32<br />

digital, 15<br />

dilation, 26<br />

equation, 32<br />

Dirac delta, 18–20, 59<br />

direct feedthrough, 21<br />

discrete<br />

149


150 INDEX<br />

Fourier series, 17<br />

Fourier transform, 17<br />

time, 15<br />

wavelet transform, 27<br />

domain, 15<br />

double-shift orthogonality, 30<br />

downsampling, 28<br />

dyadic points, 32<br />

dyadic scales, 31<br />

dynamic translinear circuits, 43<br />

dynamical matrix, 21<br />

ECG, 10<br />

intracardiac, 10<br />

subcutaneous, 10<br />

surface, 10<br />

electrical systole, 13<br />

electrocardiogram, 10<br />

elementary J-inner factors, 93<br />

energy distribution, 73<br />

Euler’s formula, 16<br />

even phase, 35<br />

exponent<br />

Hölder, 34<br />

Lipschitz, 34<br />

fibrillation<br />

atrial, 14<br />

ventricular, 14<br />

filter bank, 27, 72<br />

filter synthesis, 48<br />

final value theorem, 19<br />

finite energy, 16<br />

finite impulse response, 23, 29, 89<br />

Fourier<br />

series, 16<br />

discrete, 17<br />

transform, 15<br />

continuous, 16<br />

discrete, 17<br />

windowed, 18<br />

frequency domain, 15<br />

frequency response function, 22<br />

full rate, 34<br />

function, vector, 40<br />

Gabor transform, 18<br />

Gaussian wavelet, 27<br />

global optimization, 51, 57, 86<br />

Hölder exponent, 34<br />

Haar wavelet, 28<br />

Hadamard product, 123<br />

half rate, 35<br />

Halmos extension, 95<br />

Hankel matrix, 61<br />

Hankel singular values, 60<br />

heart, 8<br />

Heisenberg uncertainty rectangle, 26<br />

Hermitian transpose, 42<br />

high-pass filter, 28<br />

homomorphic filtering, 119<br />

implantable devices, 7, 43<br />

impulse<br />

continuous-time, 18, 20<br />

discrete-time, 18<br />

response function, 20<br />

impulse response, 45<br />

indicator function, 54<br />

initial phase angles, 16<br />

initial value theorem, 18<br />

inner-product, 51<br />

input matrix, 21<br />

input vector, 21<br />

integral wavelet transform, 26<br />

intracardiac ECG, 10<br />

iteration scheme, 32<br />

Kronecker delta, 18, 20, 59<br />

L 2 -approximation, 51<br />

Laplace transform, 18<br />

lattice form, 75<br />

lead, 10<br />

linear fractional transform, 93<br />

linear phase, 22, 34<br />

linear system, 19<br />

time-invariant, 19<br />

Lipschitz exponent, 34<br />

local optima, 82<br />

lossless systems, 41, 89, 90<br />

low-pass filter, 28<br />

Lyapunov-Stein equations, 60<br />

magnetic resonance imaging, 118<br />

Mallat’s algorithm, 31<br />

matrix exponential, 53<br />

McMillan degree, 20<br />

Mexican Hat wavelet, 27<br />

modulation matrix, 29<br />

mother wavelet, 34<br />

moving average filters, 23<br />

multi-resolution, 25, 31<br />

multiplicative noise, 119<br />

multiwavelet, 34, 40, 89<br />

design, 89<br />

parameterization, 42


INDEX 151<br />

myocardium, 8<br />

normalized convolution, 123<br />

nuclear magnetic resonance imaging, 118<br />

Nyquist frequency, 15<br />

Nyquist rate, 15<br />

observability grammian, 61<br />

observability matrix, 61<br />

odd phase, 35<br />

orthogonal complement, 32<br />

orthogonal filter banks, 29<br />

orthogonal wavelets, 75<br />

orthogonality condition, 29<br />

orthonormal wavelets, 25<br />

output matrix, 21<br />

output vector, 21<br />

overcomplete wavelet transform, 37<br />

P wave, 12<br />

pacemaker, 8<br />

pacemaker cells, 9<br />

Padé approximation, 45<br />

Padé approximation, 48<br />

Parseval’s identity, 51<br />

perfect reconstruction, 29<br />

phantom, 118<br />

phase, 34<br />

polarized, 8<br />

poles, 20<br />

polyphase filters, 34, 75, 76<br />

polyphase matrix, 35<br />

power, 16<br />

power complementary, 29<br />

PP interval, 13<br />

PQ segment, 12<br />

proper rational transfer function, 20<br />

pseudofrequency, 26<br />

QRS complex, 12<br />

QT interval, 13, 114<br />

quadrature mirror filters, 30<br />

radio frequency inhomogeneity, 118<br />

reduced rate, 35<br />

region of convergence, 18<br />

regularity, 34<br />

repolarization, 9, 12<br />

resting potential, 8<br />

rhythm, 14<br />

rotation matrix, 76<br />

RR interval, 13<br />

sampling<br />

frequency, 15<br />

rate, 15<br />

scaling<br />

coefficients, 28<br />

filter coefficients, 30<br />

function, 31<br />

Schur<br />

form, 98<br />

tangential algorithm, 93<br />

vector, 94<br />

segments (ECG), 12<br />

separation of exponentials, 54<br />

short time Fourier transform, 18<br />

signal processing, 7<br />

similarity transformation, 21<br />

sino-atrial node, 9<br />

sinusoidal fidelity, 22<br />

Smith-Barnwell orthogonality conditions, 40<br />

space, approximation, 31<br />

space,detail, 32<br />

sparsity, 72<br />

ST segment, 12<br />

stability, 50, 54, 90<br />

starting point, 57<br />

state matrix, 21<br />

state vector, 21<br />

state-space, 20<br />

dimension, 21<br />

representation, 21, 54<br />

stationary wavelet transform, 37<br />

stationary wavelet transform, polyphase, 38<br />

step response function, 20<br />

strictly proper rational function, 45<br />

subcutaneous ECG, 10<br />

sudden cardiac death, 14<br />

surface ECG, 10<br />

syncytium, 14<br />

synthesis, 16<br />

system matrix, 21<br />

T wave, 12, 114<br />

tangential Schur algorithm, 93<br />

Taylor series expansion, 48<br />

time domain, 15<br />

time-frequency localization, 25<br />

time-invariant systems, 19<br />

time-shifting, 26, 47<br />

transfer function, 20<br />

proper rational, 20<br />

transform<br />

Fourier, 15<br />

continuous, 16<br />

discrete, 17


152 INDEX<br />

windowed, 18<br />

Gabor, 18<br />

Laplace, 18<br />

wavelet, 25<br />

continuous, 26<br />

discrete, 27<br />

truncation error, 47<br />

uncertainty rectangle, 18, 26<br />

upsampling, 28<br />

vanishing moments, 33, 55, 79, 80, 101<br />

variance maximization, 73<br />

vector function, 40<br />

vectorfunctions, 40<br />

ventricle, 8<br />

ventricular fibrillation, 14<br />

wavelet, 71<br />

approximation, 45<br />

basis, 33, 71<br />

choosing a, 72<br />

coefficients, 28<br />

design, 72<br />

equation, 32<br />

filter coefficients, 30<br />

function, 31, 32<br />

Gaussian, 27<br />

Mexican Hat, 27<br />

multi, 40, 89<br />

orthonormal, 25, 75<br />

transform, 25<br />

continuous, 26<br />

discrete, 27<br />

overcomplete, 37<br />

stationary, 37<br />

wavelets<br />

Daubechies, 34<br />

waves (ECG), 12<br />

windowed Fourier transform, 18<br />

zero-order hold, 18, 59<br />

zeros, 20<br />

zoom-in property, 26

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!