01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

69<br />

Figure 4-2: Illustration of the fact that in a probabilistic regressive model the uncertainty on the predictive<br />

model increases quadratically with the amplifude of the input.<br />

It also grows with the number of training points. This effect can be diminished if one uses a<br />

sparse representation of the data. Parametric models, such as Gaussian Mixture Regression<br />

which we will see next, can also remove the requirement to store all of the training points. Here<br />

we considered solely linear regressive models, non-linear probabilistic regressive models, such<br />

as Gaussian Process Regression, will be covered in Section 5.8.<br />

4.4 Gaussian Mixture Regression<br />

Adapted from Hsi Guang Sung, Gaussian Mixture Regression and Classification, PhD thesis, Rice University,<br />

2004.<br />

In Section 3.1.5, we introduced Gaussian Mixture Models (GMM). Assume<br />

N<br />

z ∈° a multi-<br />

p z = p z ,..., 1<br />

z of the joint distribution of x<br />

N<br />

= 1,.., K Gaussians ( | , ) ( , )<br />

k k<br />

k k<br />

k N<br />

pk<br />

z µ ∑ = N ⎡⎡<br />

⎣⎣µ<br />

∑ ⎤⎤ with mean<br />

⎦⎦ µ ∈°<br />

k N×<br />

N<br />

∑ ∈° , i.e.<br />

dimensional variable GMM builds a model of ( ) ( )<br />

using a mixture of k<br />

and covariance matrix<br />

\<br />

( )<br />

1<br />

p z p z z z<br />

k= 1 k=<br />

1 k<br />

2π<br />

∑<br />

2<br />

K<br />

K<br />

1<br />

k k k<br />

T<br />

k<br />

−<br />

k<br />

( ) = ∑α k<br />

⋅ k ( | µ , ∑ ) = ∑ α<br />

k<br />

⋅ exp<br />

N ( −µ ) ( ∑ ) ( −µ<br />

) (4.18)<br />

The mixing coefficientsα satisfy<br />

k ∑ αk<br />

= 1.<br />

K<br />

k = 1<br />

Gaussian Mixture Regression (GMR) exploits the joint density to construct its estimate using the<br />

expectation on the conditional of the variable onto which we wish to make a prediction. If for<br />

instance, we wish to predict z1<br />

from knowing all other entries z z , then we can compute:<br />

2<br />

,... N<br />

{ ( N)<br />

} ( N)<br />

∫<br />

(4.19)<br />

zˆ = E p z | z ,..., z = z ⋅p z | z ,..., z ⋅dz<br />

1 1 2 1 1 2 1<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!