MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
184<br />
9.1.2 Singular Value Decomposition (SVD)<br />
When the eigenvectors are not linearly independent, then X does not have an inverse (it is thus<br />
singular) and such decomposition does not exist. The eigenvalue decomposition consists then in<br />
finding a similarity transformation such that:<br />
A= UΛ V<br />
(8.11)<br />
With UV , two orthogonal (if real) or unitary (if complex) matrices and Λ a diagonal matrix. Such<br />
decomposition is called singular value decomposition (SVD). SVD are useful in sofar that A<br />
represents the mapping of a n-dimensional space onto itself, where n is the dimension of A.<br />
#<br />
An alternative to SVD is to compute the Moore-Penrose Pseudoinverse A of the non invertible<br />
#<br />
matrix A and then exploit the fact that, for a pair of vectors z and c, z = A c is the shortest length<br />
least-square solution to the problem Az = c . Methods such as PCA that find the optimal (in a<br />
least-square sense) projection of a dataset can be approximated using the pseudoinverse when<br />
the transformation matrix is singular.<br />
9.1.3 Frobenius Norm<br />
The Frobenius norm of an m× nmatrix A is given by:<br />
A<br />
F<br />
m<br />
n<br />
i= 1 j=<br />
1<br />
ij<br />
2<br />
= ∑∑ a<br />
(8.12)<br />
9.2 Recall of basic notions of statistics and probabilities<br />
9.2.1 Probabilities<br />
Consider two variables x and y taking discrete values over the intervals [x 1 ,…, x M ] and [y 1 ,…, y N ]<br />
respectively, then P( x=<br />
x i ) is the probability that the variable x takes value x i<br />
, with:<br />
i) 0 ≤ P( x= x ) ≤1, ∀ i=<br />
1,..., M,<br />
M<br />
∑<br />
i=<br />
1<br />
( x )<br />
ii) P x= = 1.<br />
i<br />
i<br />
( j )<br />
The same two above properties applies to the probabilities P y= y , ∀ j=<br />
1,... N.<br />
Some properties follow from the above:<br />
Let P( x= a)<br />
be the probability that the variable x will take the value a. If P( x a) 1<br />
= = , x is a<br />
constant with value a. If x is an integer and can take any value a between [1, N]<br />
∈• with equal<br />
1<br />
probability, then the probability that x takes value a is Px ( = a)<br />
=<br />
N<br />
© A.G.Billard 2004 – Last Update March 2011