MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
160<br />
7 Markov-Based Models<br />
Except for ICA, recurrent neural networks and continuous Hopfield network, these lectures notes<br />
have covered primarily methods to encode static datasets, i.e. data that do not vary in time. Being<br />
able to predict patterns that vary in time is fundamental to many applications. Such patterns are<br />
often referred to as time-series.<br />
In this chapter, we cover two models to encode time-series based on the hypothesis of a firstorder<br />
Markov Process, which we describe next.<br />
7.1 Markov Process<br />
A Markov Process is by definition discrete and assumes that the evolution of a system can be<br />
described by a finite set of states. If<br />
N particular values { s ,... 1<br />
s<br />
N}<br />
, i.e. x { s ,... 1<br />
sN}<br />
x is the state of the system at time, then x can take only<br />
t<br />
∈ .<br />
A first-order markov process is a process by which the future state of the system can be<br />
determined solely by knowing the current state of the system. In other words, if one has made T<br />
consecutive observation of the state x of the system: { t} t 1<br />
the system at the next time step<br />
T 1<br />
current state, i.e.:<br />
T<br />
X x =<br />
= , the probability that the state of<br />
x +<br />
take any of the N possible value is determined solely by the<br />
( t+ 1 j<br />
|<br />
t, t− 1,...., 0) ( t+<br />
1 j<br />
|<br />
t)<br />
p x = s x x x = p x = s x<br />
(7.1)<br />
When modeling time-series, to assume that a process description is first-order markov is<br />
advantageous in that it simplifies greatly computation. Instead of computing the complete<br />
x , x ,...., x<br />
conditional on all previously observed states , one must solely compute the<br />
t t−<br />
1 0<br />
conditional on the last state. We will see how this is exploited in two different techniques widely<br />
used in machine learning next.<br />
© A.G.Billard 2004 – Last Update March 2011