MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
162<br />
Figure 7-1: Schematic illustrating the concepts of a HMM. The process is assumed to be composed of 7<br />
hidden states. All transitions across all states are possible (fully connected). Note that for simplicity, we<br />
show only arrows across adjacent states. The schematic at the bottom shows a particular instance of<br />
transition across states over five time steps. Transitiosn across each state leads to an observation of a<br />
particular set of values for the system’s variables. As we see, the system rotates first across states 1 and 2<br />
and then stays for two time steps on state 3.<br />
Transitions across states are described as a stochastic finite automata process. The stochasticity<br />
of the process is represented by computing a set of transition probabilities that determine the<br />
likelihood to stay in a state or to jump to another state. Transition probabilities are encapsulated<br />
in a N N<br />
represent the probability of transiting from<br />
× matrix A , whose elements { } ij<br />
i , j=<br />
1...<br />
state i to state j , i.e. aij p( sj | si<br />
)<br />
a<br />
N<br />
= . The sum of all elements in each row of A equals 1. Each<br />
state is associated an initial probability π<br />
i, i= 1,.. Nthat represents the likelihood to be in that<br />
state at any given point in time. In addition, one assigns to each state i a density ( ) i<br />
b o , socalled<br />
the emission probability that determines the probability of the observation to take a<br />
particular value when in state S . Depending on whether the observables take discrete versus<br />
i<br />
continuous values, we talk about a discrete versus continuous HMM. When continuous, the<br />
density may be estimated through a Gaussian Mixture Model. In this case, one associates a<br />
GMM per state. HMM are used widely in speech processing. There, often, one uses very few<br />
hidden states (at maximum 3!). The complexity of the speech is then embedded in the GMM<br />
density modeling associated at each state.<br />
© A.G.Billard 2004 – Last Update March 2011