13.08.2022 Views

advanced-algorithmic-trading

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

273

p∑

K(x i , x k ) = (1 + x ij x kj ) d (19.20)

This provides a significantly more flexible decision boundary and essentially amounts to fitting

a SVC in a higher-dimensional feature space involving d-degree polynomials of the features (see

Figure 19.10).

j=1

Figure 19.10: A d-degree polynomial kernel; A radial kernel

Hence, the definition of a support vector machine is a support vector classifier with a nonlinear

kernel function.

We can also consider the popular radial kernel (see Figure 19.10):

K(x i , x k ) = exp ⎝−γ

p∑

(x ij − x kj ) 2 ⎠ , γ > 0 (19.21)

j=1

So how do radial kernels work? They clearly differ from polynomial kernels. Essentially if our

test observation x ∗ is far from a training observation x i in standard Euclidean distance then the

sum ∑ p

j=1 (x∗ j − x ij) 2 will be large and thus K(x ∗ , x i ) will be very small. Hence this particular

training observation x i will have almost no effect on where the test observation x ∗ is placed, via

f(x ∗ ).

Thus the radial kernel has extremely localised behaviour and only nearby training observations

to x ∗ will have an impact on its class label.

19.8.1 Biblographic Notes

Originally, SVMs were invented by Vapnik[98], while the current standard "soft margin" approach

is due to Cortes[34]. My treatment of the material follows, and is strongly influenced by, the

excellent statistical machine learning texts of James et al[59] and Hastie et al[51].

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!