Segmentation of heterogeneous document images : an ... - Tel

Segmentation of heterogeneous document images : an ... - Tel Segmentation of heterogeneous document images : an ... - Tel

tel.archives.ouvertes.fr
from tel.archives.ouvertes.fr More from this publisher
14.01.2014 Views

tel-00912566, version 1 - 2 Dec 2013 Figure 4.16: This figure shows the evolution of training error rate with 5 different learning rates. All other parameters remain the same. obtained error in each iteration. Any other parameter of the training remains the same unless if stated otherwise. Learning rate Learning rate refers to the learning rate of the voted perceptron training algorithm. It controls the fraction of computed deltas in each iteration of the training that are contributing to the weights of the feature functions. The number of ICM cycles is 10. The block overlap ratio is 0.25 and the width and height of each block is half of the mean width and height of all text characters of the page, respectively. Figure 4.16 shows the evolution of training error with 5 different learning rates. It suggests that a high value for learning rate forces the training to go into a cycle. On the other hand, the training process is smooth with a small value for learning rate (0.1 in this case) however the process converges to a higher training error. In this particular experiment, a learning rate of 0.25 can be considered as good because the number of misclassified sites decreases reasonably fast without going into cycles. Unfortunately, there are no rules of thumb to determine a best value for learning rate without considering the number of blocks, number of features functions and their characteristics. 84

tel-00912566, version 1 - 2 Dec 2013 Figure 4.17: This figure shows the evolution of training error rate with 3 different overlapping ratios. Overlapping ratio Overlapping ratio refers to the amount that two consecutive blocks overlap. For this experiment we consider three different overlapping ratio; without overlap, with 25% and 50% overlap. Everything else remain the same but the total number of sites. Figure 4.17 displays the percentage of misclassified sites during the training process with 3 different overlapping ratios. In conclusion, a smaller overlapping ratio results in less number of sites, much faster training process and less number of misclassified sites. However, despite the decrease in the number of misclassified sites, the number of misclassified sites that are located in between columns of text increases. For this reason, it is in our best interest to use a value higher than 0.25 for the overlapping ratio. Values higher than 0.5 result in a huge number of sites and increase the memory consumption and training time substantially. Maximum number of ICM cycles Maximum number of ICM cycles refers to the maximum number of cycles that are allowed for the iterated conditional modes inference algorithm before the algorithm converges. The ICM algorithm is supposed to converge to a fixed state of the system after several cycles, however the convergence is not guaranteed. In the former case, a maximum number of cycles is set to terminate the inference algorithm prematurely. Figure 4.18 shows the progress of the training algorithm with 5 different maximum number of ICM cycles. The results indicate that except of the first experiment with only one cycle, all other experiments perform the same way with slight changes. In conclusion 5 cycles can be considered a good value as a 85

tel-00912566, version 1 - 2 Dec 2013<br />

Figure 4.16: This figure shows the evolution <strong>of</strong> training error rate with 5 different<br />

learning rates. All other parameters remain the same.<br />

obtained error in each iteration. Any other parameter <strong>of</strong> the training remains<br />

the same unless if stated otherwise.<br />

Learning rate<br />

Learning rate refers to the learning rate <strong>of</strong> the voted perceptron training algorithm.<br />

It controls the fraction <strong>of</strong> computed deltas in each iteration <strong>of</strong> the<br />

training that are contributing to the weights <strong>of</strong> the feature functions. The number<br />

<strong>of</strong> ICM cycles is 10. The block overlap ratio is 0.25 <strong>an</strong>d the width <strong>an</strong>d height<br />

<strong>of</strong> each block is half <strong>of</strong> the me<strong>an</strong> width <strong>an</strong>d height <strong>of</strong> all text characters <strong>of</strong> the<br />

page, respectively.<br />

Figure 4.16 shows the evolution <strong>of</strong> training error with 5 different learning<br />

rates. It suggests that a high value for learning rate forces the training to go into<br />

a cycle. On the other h<strong>an</strong>d, the training process is smooth with a small value<br />

for learning rate (0.1 in this case) however the process converges to a higher<br />

training error. In this particular experiment, a learning rate <strong>of</strong> 0.25 c<strong>an</strong> be considered<br />

as good because the number <strong>of</strong> misclassified sites decreases reasonably<br />

fast without going into cycles.<br />

Unfortunately, there are no rules <strong>of</strong> thumb to determine a best value for<br />

learning rate without considering the number <strong>of</strong> blocks, number <strong>of</strong> features<br />

functions <strong>an</strong>d their characteristics.<br />

84

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!