01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

70<br />

When applying Bayes’ theorem, we get:<br />

z<br />

1<br />

=<br />

∫<br />

∫<br />

( , ,...,<br />

N )<br />

( , ,..., ) ⋅<br />

z ⋅p z z z ⋅dz<br />

1 1 2 1<br />

p z z z dz<br />

1 2 N 1<br />

(4.20)<br />

In contrast to most regression techniques, GMR also allows to make prediction on multi-<br />

y z , z , z x= z z , then we can<br />

dimensional output in one swipe. So, if we set = { } and { }<br />

compute:<br />

1 2 3<br />

{ ( )} ( 1 2 3 4 5 )<br />

{ }<br />

4 ,..., N<br />

yˆ = E p y| x = E p z , z , z | z , z ,..., z<br />

(4.21)<br />

N<br />

4.4.1 One Gaussian Case<br />

Let us first consider first the case where the joint density p( z) p( y,<br />

x)<br />

= follows a single<br />

Gaussian distribution (and not a mixture). Using the fact that, if the joint distribution of two<br />

variables is Gaussian then each conditional distribution is also Gaussian, see Figure 9-2 for an<br />

illustration, we can compute the conditional density ( | )<br />

where µ Y<br />

and<br />

X<br />

p y x which is given by:<br />

1 −1<br />

( y YX XX x YX YX XX XY)<br />

−<br />

( | ) µ ( ) ( µ ),<br />

( )<br />

p y x = N +∑ ∑ x− ∑ −∑ ∑ ∑ (4.22)<br />

µ are the means of the variable y and x respectively and have thus the<br />

Y X<br />

dimensions of y and x , which we shall call N , N . ∑ XX<br />

variable x has<br />

dimension<br />

dimension ,<br />

N<br />

Y X X Y<br />

N × N N × N respectively.<br />

the covariance matrix of the input<br />

X X<br />

× N and the cross-covariances matrices<br />

YX ,<br />

∑ ∑ have<br />

In predictive regression, we are interested in computing the expectation and uncertainty of the<br />

predictive regressive model at a given query point. When the conditional density is Gaussian, this<br />

*<br />

is easy to obtain and we thus have for a query point x , the estimated output:<br />

−1<br />

{ ( )} µ ( ) ( µ )<br />

* *<br />

yˆ E p y| x<br />

Y YX XX<br />

x<br />

X<br />

= = +∑ ∑ − (4.23)<br />

XY<br />

with associated uncertainty:<br />

−<br />

( p( y x)<br />

) ( ) 1<br />

var | =∑YX −∑YX ∑XX ∑ (4.24)<br />

XY<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!