20.01.2013 Views

Master Thesis - Department of Computer Science

Master Thesis - Department of Computer Science

Master Thesis - Department of Computer Science

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

• Product Rule: Product Rule computes the s<strong>of</strong>t class label vectors as:<br />

˜d j (x) = d j<br />

Null (x) ∗ djRange<br />

(x), j = 1, ...., C (4.33)<br />

4.4.2 Proposed Technique for Decision Fusion<br />

Unlike sum rule and product rule, our technique <strong>of</strong> decision fusion learns each <strong>of</strong><br />

the classifier’s behavior by using training data at decision level. Learning a classifier<br />

involves training with the response <strong>of</strong> that classifier. So the response vectors on a<br />

validation set <strong>of</strong> a database which is disjoint from training and testing set can be<br />

used as training information at decision level for learning the behavior <strong>of</strong> a classifier.<br />

We employ LDA and nonparametric LDA on training information at decision level to<br />

build up discriminative eigenmodel. Then response vectors <strong>of</strong> the input test images<br />

are projected in that eigenmodel and for each test response vector, the similarity<br />

measures for the respective classes in the eigenmodel provide a new response vector<br />

which replaces the old test response vector.<br />

To accomplish the whole task, we divide our database into three disjoint subsets,<br />

training, validation and testing sets. Now the training set along with validation set<br />

and testing set respectively constructs the “training response vector set” and “testing<br />

response vector set”, respectively at classifier output space. Then “training response<br />

vector set” is used as training data to construct a LDA or nonparametric LDA-based<br />

eigenmodel at classifier response level.<br />

Here, we formally define validation and testing set which are independent <strong>of</strong> any<br />

classifier. Let, the validation and testing set <strong>of</strong> a database are denoted by XV A and<br />

XT S and can be defined as,<br />

XV A = [vx 1 1 , vx12 , . . . , vx1V , vx21 , . . . , vxij , . . . , vxCV ], (4.34)<br />

XT S = [tx 1 1, tx 1 2, . . . , tx 1 S, tx 2 1, . . . , tx i j, . . . , tx C S ]. (4.35)<br />

where vx i j represents the jth validation sample from class i and V is the number <strong>of</strong><br />

validation samples per class. Similarly, j th testing sample from class i is denoted<br />

by tx i j<br />

and number <strong>of</strong> testing samples per class is represented as S. The “training<br />

response vector set” and “testing response vector set” for null space classifier are<br />

84

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!