Segmentation of heterogeneous document images : an ... - Tel
Segmentation of heterogeneous document images : an ... - Tel
Segmentation of heterogeneous document images : an ... - Tel
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
would be difficult to draw a conclusion solely based on the reported accuracy<br />
rate <strong>of</strong> each method. Also it is impossible to implement or find every method<br />
that is listed here. As a consequence we have to rely as much as possible on<br />
the reported accuracies when comparison is possible <strong>an</strong>d then a theoretical <strong>an</strong>d<br />
fundamental reasoning to decide which method is fit for our work. Table 2.1<br />
shows some <strong>of</strong> the characteristics <strong>of</strong> our dataset.<br />
Table 2.1:<br />
CHARACTERISTICS OF THE DOCUMENTS IN OUR CORPUS<br />
tel-00912566, version 1 - 2 Dec 2013<br />
Frequently occurs Seldom occurs Never occurs<br />
Printed text lines Sl<strong>an</strong>ted text lines Calligraphy<br />
H<strong>an</strong>dwritten text lines Vertical text lines Highly skewed text lines<br />
Large gaps between words Degraded quality<br />
Rule lines<br />
Tr<strong>an</strong>sparency effect<br />
Close text lines<br />
Multi-script <strong>document</strong>s<br />
Touching text lines Black borders<br />
Multi-column <strong>document</strong>s<br />
Salt & Paper noise<br />
Side notes<br />
Variety <strong>of</strong> font sizes<br />
In order to choose the best line detection method as the base for our work,<br />
first we have to compare methods in a situation where both datasets <strong>an</strong>d the<br />
evaluation metrics are the same, then after pruning weak methods, we have<br />
to go into details <strong>of</strong> remaining methods to reason whether they might fail to<br />
achieve good results when applied to our <strong>document</strong>s or not.<br />
We have named fourteen methods for detection <strong>of</strong> text lines so far. Table 2.2<br />
presents some <strong>of</strong> these methods, <strong>an</strong>d the accuracy rate that they have achieved<br />
by applying to ICDAR2007 [39] dataset.<br />
Table 2.2: ACCURACY RATES FOR METHODS APPLIED TO ICDAR2007<br />
DATASET [39]<br />
Method Reported In dataset Evaluation Method Accuracy<br />
Papavassiliou 2010 [75] [75] ICDAR07 Match Counting 98.3%<br />
Louloudis 2009 [58] [58] ICDAR07 Match Counting 97.4%<br />
Stafylakis 2008[90] [90] ICDAR07 Match Counting 97.1%<br />
Louloudis 2006 [56] [39] ICDAR07 Match Counting 95.4%<br />
Bukhari 2009a [18] [18] ICDAR07 Match Counting 90.7%<br />
Another table 2.3 presents the detection accuracy rate for CBDAR2007<br />
dataset. Note that for the two methods reported there, the comparison is a<br />
challenge because one method gains the upper h<strong>an</strong>d using one evaluation metric<br />
<strong>an</strong>d using <strong>an</strong>other evaluation metric the situation ch<strong>an</strong>ges. To solve this<br />
dilemma, we decide to choose the evaluation metric that is more appropriate<br />
<strong>an</strong>d well designed. In this case, it is Match Counting, which is the same evaluation<br />
that we have described in Appendix A. Based on this evaluation metric<br />
the result <strong>of</strong> both methods c<strong>an</strong> be considered the same.<br />
33