01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

162<br />

Figure 7-1: Schematic illustrating the concepts of a HMM. The process is assumed to be composed of 7<br />

hidden states. All transitions across all states are possible (fully connected). Note that for simplicity, we<br />

show only arrows across adjacent states. The schematic at the bottom shows a particular instance of<br />

transition across states over five time steps. Transitiosn across each state leads to an observation of a<br />

particular set of values for the system’s variables. As we see, the system rotates first across states 1 and 2<br />

and then stays for two time steps on state 3.<br />

Transitions across states are described as a stochastic finite automata process. The stochasticity<br />

of the process is represented by computing a set of transition probabilities that determine the<br />

likelihood to stay in a state or to jump to another state. Transition probabilities are encapsulated<br />

in a N N<br />

represent the probability of transiting from<br />

× matrix A , whose elements { } ij<br />

i , j=<br />

1...<br />

state i to state j , i.e. aij p( sj | si<br />

)<br />

a<br />

N<br />

= . The sum of all elements in each row of A equals 1. Each<br />

state is associated an initial probability π<br />

i, i= 1,.. Nthat represents the likelihood to be in that<br />

state at any given point in time. In addition, one assigns to each state i a density ( ) i<br />

b o , socalled<br />

the emission probability that determines the probability of the observation to take a<br />

particular value when in state S . Depending on whether the observables take discrete versus<br />

i<br />

continuous values, we talk about a discrete versus continuous HMM. When continuous, the<br />

density may be estimated through a Gaussian Mixture Model. In this case, one associates a<br />

GMM per state. HMM are used widely in speech processing. There, often, one uses very few<br />

hidden states (at maximum 3!). The complexity of the speech is then embedded in the GMM<br />

density modeling associated at each state.<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!