08.12.2012 Views

Journal of Software - Academy Publisher

Journal of Software - Academy Publisher

Journal of Software - Academy Publisher

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

JOURNAL OF SOFTWARE, VOL. 6, NO. 5, MAY 2011 883<br />

( n−c) ni<br />

|| xi −x||<br />

SSA ( c −1)<br />

i=<br />

1<br />

F = =<br />

c ni<br />

SSE ( n − c)<br />

( c −1) || x −x<br />

||<br />

c<br />

∑<br />

∑∑<br />

i= 1 t=<br />

1<br />

it i<br />

c<br />

2<br />

( n−c) ˆ ∑ nd i ( xi, x)<br />

i=<br />

1<br />

c− c ni<br />

2<br />

∑∑d<br />

xit xi<br />

i= 1 t=<br />

1<br />

SSA ( c −1)<br />

KF = =<br />

SSE ( n − c)<br />

( 1) ˆ ( , )<br />

2<br />

2<br />

(11)<br />

(12)<br />

Where: KF is statistic <strong>of</strong> samples vectors in feature<br />

space, SSA is between class variance, SSE is inner-class<br />

variance, xi is the mean vector <strong>of</strong> the ith class samples<br />

vectors, x is the mean vector <strong>of</strong> the whole samples<br />

vectors, it x is the ith class and the tth sample vector, i n<br />

is amount <strong>of</strong> samples <strong>of</strong> the ith class.<br />

C. SVM Algorithm<br />

According to the above classification result, we can<br />

acquire a training set S = {( x1, y1),( x2, y2), � ,( xN,<br />

yN)}<br />

p<br />

<strong>of</strong> N data points, where i ∈Ω x is the ith input pattern<br />

and yi ∈ R is the ith output pattern. In most cases, the<br />

searching <strong>of</strong> a suitable hyperplane in an input space is too<br />

restrictive to be <strong>of</strong> practical use. Hence, suppose<br />

m<br />

{ ϕ j ( x)} j=<br />

1 represents the nonlinear transfer set from<br />

the input space to feature space, where m is the dimension<br />

<strong>of</strong> feature space. Hence, a decision hyperplane in feature<br />

can be defined as<br />

y i<br />

T<br />

i[ w ϕ ( x ) + b]<br />

≥ 1,<br />

i 1, �,<br />

N,<br />

= (13)<br />

Where ϕ (⋅)<br />

is a kernel function which maps the input<br />

space into a higher dimensional space, m and n are,<br />

respectively, the dimensions <strong>of</strong> the input space and<br />

feature space. However, this function is not explicitly<br />

constructed. In order to have the possibility to violate<br />

(15), in case a separating hyperplane in this higher<br />

dimensional space does not exist, slack variables ξ k are<br />

introduced such that<br />

T ⎧y<br />

i[<br />

w ϕ(<br />

xi<br />

) + b]<br />

≥ 1−<br />

ξi<br />

, i = 1,<br />

�,<br />

N<br />

⎨<br />

⎩ξi<br />

≥ 0,<br />

i = 1,<br />

�,<br />

N<br />

(14)<br />

Subsequently, according to the structural risk<br />

minimization principle, the risk bound is minimized by<br />

considering the optimization problem<br />

1<br />

w<br />

T<br />

Minimize ⋅ w + C∑<br />

ξ i (15)<br />

2<br />

= 1<br />

subject to (14). Where C is a constant and can be<br />

regarded as a regularization parameter. Tuning this<br />

parameter can obtain a balance between margin<br />

maximization and classification violation. In order to<br />

© 2011 ACADEMY PUBLISHER<br />

i<br />

N<br />

solve the constraint optimal problem, one constructs the<br />

Lagrangian and transformed into the dual<br />

⎧Maximize<br />

⎪<br />

N N<br />

⎪<br />

1<br />

i= 1, �,<br />

N<br />

⎪W(<br />

α) = ∑αi − ∑αα<br />

i iyiyK i ( xi, xj)<br />

⎨<br />

i= 1 2 i, j=<br />

1<br />

⎪<br />

N<br />

⎪ Subject to ∑ yiαi = 0, 0 ≤αi ≤ C, i = 1, �,<br />

N<br />

⎪⎩<br />

i=<br />

1<br />

(16)<br />

Searching the optimal hyperplane in (15) is a quadratic<br />

programming (QP) problem, according to the Kuhn-<br />

Tucker theorem, the solution <strong>of</strong> the optimal problem must<br />

satisfies the equality<br />

⎧αi(<br />

yi( xw i i + b)<br />

− 1 + ξi)<br />

= 0<br />

⎨<br />

⎩(<br />

C − αi) ξi<br />

= 0<br />

, i = 1, �,<br />

N (17)<br />

In (12), α i is zero for most <strong>of</strong> samples, when the non-<br />

zero values α i are satisfied with the equality sign in<br />

(14), the pattern x i corresponding with α i is called<br />

support vector.<br />

*<br />

If α is the optimal solution in (16), then<br />

w<br />

*<br />

=<br />

N<br />

*<br />

∑ α i yi<br />

xi<br />

(18)<br />

i=<br />

1<br />

That is the weight coefficient vector <strong>of</strong> optimal<br />

classification hyperplane is linear combination <strong>of</strong> training<br />

pattern vectors. After solving the above problems, the<br />

optimal classification function can be acquired as<br />

N<br />

⎛ *<br />

* ⎞<br />

f ( x)<br />

= sign⎜∑<br />

yiα<br />

i K(<br />

xi<br />

⋅ x)<br />

+ b ⎟ (19)<br />

⎝ i=<br />

1<br />

⎠<br />

IV. RESULET AND DISCUSSION<br />

In order to evaluate the performance <strong>of</strong> BSE and verify<br />

effect <strong>of</strong> these algorithms, the improved FKCM and SVM<br />

were performed on IRIS dataset and measurement dataset<br />

respectively. Matlab7.01 s<strong>of</strong>tware was utilized for data<br />

processing and analyzing.<br />

A. Pre-processing <strong>of</strong> measurement dataset<br />

After data normalization, procedures <strong>of</strong> correlation<br />

analysis, weight computation, and initial clustering center<br />

were performed on IRIS dataset and measurement dataset<br />

in turn. Correlation coefficients <strong>of</strong> IRIS dataset were<br />

shown in Tab. II and Tab. III by using (2). However,<br />

some indexes with strong correlation can be eliminated<br />

by Pearson Correlation coefficient analysis method.<br />

Because <strong>of</strong> strong correlation in indexes, 4 z , 5 z and 4 z′ can<br />

be deleted in data analysis process. Furthermore, index<br />

weight coefficient <strong>of</strong> measurement dataset and IRIS<br />

dataset was, respectively, WM = (0.2399 0.2858 0.2967<br />

0.1776) and WIRIS = (0.2374 0.1785 0.2785 0.30) by<br />

using (3).<br />

FKCM and SVM Performed on IRIS Dataset

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!