01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

160<br />

7 Markov-Based Models<br />

Except for ICA, recurrent neural networks and continuous Hopfield network, these lectures notes<br />

have covered primarily methods to encode static datasets, i.e. data that do not vary in time. Being<br />

able to predict patterns that vary in time is fundamental to many applications. Such patterns are<br />

often referred to as time-series.<br />

In this chapter, we cover two models to encode time-series based on the hypothesis of a firstorder<br />

Markov Process, which we describe next.<br />

7.1 Markov Process<br />

A Markov Process is by definition discrete and assumes that the evolution of a system can be<br />

described by a finite set of states. If<br />

N particular values { s ,... 1<br />

s<br />

N}<br />

, i.e. x { s ,... 1<br />

sN}<br />

x is the state of the system at time, then x can take only<br />

t<br />

∈ .<br />

A first-order markov process is a process by which the future state of the system can be<br />

determined solely by knowing the current state of the system. In other words, if one has made T<br />

consecutive observation of the state x of the system: { t} t 1<br />

the system at the next time step<br />

T 1<br />

current state, i.e.:<br />

T<br />

X x =<br />

= , the probability that the state of<br />

x +<br />

take any of the N possible value is determined solely by the<br />

( t+ 1 j<br />

|<br />

t, t− 1,...., 0) ( t+<br />

1 j<br />

|<br />

t)<br />

p x = s x x x = p x = s x<br />

(7.1)<br />

When modeling time-series, to assume that a process description is first-order markov is<br />

advantageous in that it simplifies greatly computation. Instead of computing the complete<br />

x , x ,...., x<br />

conditional on all previously observed states , one must solely compute the<br />

t t−<br />

1 0<br />

conditional on the last state. We will see how this is exploited in two different techniques widely<br />

used in machine learning next.<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!