01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

184<br />

9.1.2 Singular Value Decomposition (SVD)<br />

When the eigenvectors are not linearly independent, then X does not have an inverse (it is thus<br />

singular) and such decomposition does not exist. The eigenvalue decomposition consists then in<br />

finding a similarity transformation such that:<br />

A= UΛ V<br />

(8.11)<br />

With UV , two orthogonal (if real) or unitary (if complex) matrices and Λ a diagonal matrix. Such<br />

decomposition is called singular value decomposition (SVD). SVD are useful in sofar that A<br />

represents the mapping of a n-dimensional space onto itself, where n is the dimension of A.<br />

#<br />

An alternative to SVD is to compute the Moore-Penrose Pseudoinverse A of the non invertible<br />

#<br />

matrix A and then exploit the fact that, for a pair of vectors z and c, z = A c is the shortest length<br />

least-square solution to the problem Az = c . Methods such as PCA that find the optimal (in a<br />

least-square sense) projection of a dataset can be approximated using the pseudoinverse when<br />

the transformation matrix is singular.<br />

9.1.3 Frobenius Norm<br />

The Frobenius norm of an m× nmatrix A is given by:<br />

A<br />

F<br />

m<br />

n<br />

i= 1 j=<br />

1<br />

ij<br />

2<br />

= ∑∑ a<br />

(8.12)<br />

9.2 Recall of basic notions of statistics and probabilities<br />

9.2.1 Probabilities<br />

Consider two variables x and y taking discrete values over the intervals [x 1 ,…, x M ] and [y 1 ,…, y N ]<br />

respectively, then P( x=<br />

x i ) is the probability that the variable x takes value x i<br />

, with:<br />

i) 0 ≤ P( x= x ) ≤1, ∀ i=<br />

1,..., M,<br />

M<br />

∑<br />

i=<br />

1<br />

( x )<br />

ii) P x= = 1.<br />

i<br />

i<br />

( j )<br />

The same two above properties applies to the probabilities P y= y , ∀ j=<br />

1,... N.<br />

Some properties follow from the above:<br />

Let P( x= a)<br />

be the probability that the variable x will take the value a. If P( x a) 1<br />

= = , x is a<br />

constant with value a. If x is an integer and can take any value a between [1, N]<br />

∈• with equal<br />

1<br />

probability, then the probability that x takes value a is Px ( = a)<br />

=<br />

N<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!