MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
69<br />
Figure 4-2: Illustration of the fact that in a probabilistic regressive model the uncertainty on the predictive<br />
model increases quadratically with the amplifude of the input.<br />
It also grows with the number of training points. This effect can be diminished if one uses a<br />
sparse representation of the data. Parametric models, such as Gaussian Mixture Regression<br />
which we will see next, can also remove the requirement to store all of the training points. Here<br />
we considered solely linear regressive models, non-linear probabilistic regressive models, such<br />
as Gaussian Process Regression, will be covered in Section 5.8.<br />
4.4 Gaussian Mixture Regression<br />
Adapted from Hsi Guang Sung, Gaussian Mixture Regression and Classification, PhD thesis, Rice University,<br />
2004.<br />
In Section 3.1.5, we introduced Gaussian Mixture Models (GMM). Assume<br />
N<br />
z ∈° a multi-<br />
p z = p z ,..., 1<br />
z of the joint distribution of x<br />
N<br />
= 1,.., K Gaussians ( | , ) ( , )<br />
k k<br />
k k<br />
k N<br />
pk<br />
z µ ∑ = N ⎡⎡<br />
⎣⎣µ<br />
∑ ⎤⎤ with mean<br />
⎦⎦ µ ∈°<br />
k N×<br />
N<br />
∑ ∈° , i.e.<br />
dimensional variable GMM builds a model of ( ) ( )<br />
using a mixture of k<br />
and covariance matrix<br />
\<br />
( )<br />
1<br />
p z p z z z<br />
k= 1 k=<br />
1 k<br />
2π<br />
∑<br />
2<br />
K<br />
K<br />
1<br />
k k k<br />
T<br />
k<br />
−<br />
k<br />
( ) = ∑α k<br />
⋅ k ( | µ , ∑ ) = ∑ α<br />
k<br />
⋅ exp<br />
N ( −µ ) ( ∑ ) ( −µ<br />
) (4.18)<br />
The mixing coefficientsα satisfy<br />
k ∑ αk<br />
= 1.<br />
K<br />
k = 1<br />
Gaussian Mixture Regression (GMR) exploits the joint density to construct its estimate using the<br />
expectation on the conditional of the variable onto which we wish to make a prediction. If for<br />
instance, we wish to predict z1<br />
from knowing all other entries z z , then we can compute:<br />
2<br />
,... N<br />
{ ( N)<br />
} ( N)<br />
∫<br />
(4.19)<br />
zˆ = E p z | z ,..., z = z ⋅p z | z ,..., z ⋅dz<br />
1 1 2 1 1 2 1<br />
© A.G.Billard 2004 – Last Update March 2011