MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
MACHINE LEARNING TECHNIQUES - LASA
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
83<br />
additional ridge parameter induces a beneficial control of over-fitting and enhances the numerical<br />
stability of the solutions. (Bach & Jordan, 2002) report that, in many experiments, the solution of<br />
this regularized problem show better generalization abilities than the kernel canonical vectors, in<br />
the sense of giving higher correlated scores for new objects.<br />
Generalizing to multiple multi-dimensional datasets:<br />
As in linear CCA, one can extend this to comparison across several multidimensional datasets. If<br />
may however have different dimensions<br />
X ,...., 1<br />
X are the L datasets, each of which are composed of M observations. Each dataset<br />
L<br />
N ,.... 1<br />
N and hence<br />
L<br />
Xi<br />
: Ni× M . As in the two-<br />
K K and the solution to this<br />
dimensional case, one can construct a set of L Gram matrices ,....., 1 L<br />
kernel CCA problem is given by solving:<br />
2<br />
⎛⎛⎛⎛<br />
2 Mκ<br />
⎞⎞<br />
⎞⎞<br />
⎜⎜ K1<br />
+ I 0<br />
0 KK<br />
1 2<br />
....... KK<br />
1 L<br />
α<br />
⎜⎜ ⎟⎟<br />
⎟⎟<br />
⎛⎛ ⎞⎞⎛⎛ 1 ⎞⎞ ⎜⎜⎝⎝ 2 ⎠⎠<br />
⎟⎟ ⎛⎛ α1<br />
⎞⎞<br />
⎜⎜ ⎟⎟⎜⎜ ⎟⎟<br />
KK<br />
2 1<br />
0 ........ KK<br />
2 L<br />
α ⎜⎜ ⎟⎟ ⎜⎜ ⎟⎟<br />
⎜⎜ ⎟⎟⎜⎜ 2 ⎟⎟<br />
α2<br />
⎜⎜ ⎟⎟ ⎜⎜ ⎟⎟<br />
⎜⎜ . ⎟⎟⎜⎜ . ⎟⎟= ρ ⎜⎜.......<br />
......................................... ⎜⎜ . ⎟⎟ .<br />
⎜⎜ ⎟⎟⎜⎜ ⎟⎟<br />
⎜⎜ . ⎟⎟⎜⎜ . ⎟⎟<br />
⎜⎜ ⎟⎟ ⎜⎜ ⎟⎟<br />
.<br />
⎜⎜ ⎟⎟ ⎟⎟ ⎜⎜ ⎟⎟<br />
⎜⎜KK L 1<br />
KK<br />
L 2<br />
....... KK ⎟⎟⎜⎜<br />
L L<br />
α ⎟⎟<br />
2<br />
⎝⎝ ⎠⎠⎝⎝ L⎠⎠<br />
⎜⎜ ⎜⎜<br />
2 Mκ<br />
α ⎟⎟<br />
⎛⎛ ⎞⎞ ⎟⎟ ⎝⎝ L ⎠⎠<br />
⎜⎜ 0 ⎜⎜KL<br />
+ I ⎟⎟ ⎟⎟<br />
⎝⎝<br />
⎝⎝ 2 ⎠⎠ ⎠⎠<br />
(5.19)<br />
When using a Gaussian kernel, one can see that as the kernel width s increases (or decreases if<br />
s