deep arabic document layout analysis€¦ · ibrahim m. amer, salma hamdy and mostafa g. m. mostafa...

9
Deep Arabic Document Layout Analysis Ibrahim M. Amer, Salma Hamdy and Mostafa G. M. Mostafa This is the accepted version of a conference paper published by IEEE. Amer, I.M., Hamdy, S. & Mostafa, M.G.M. (2017). Deep Arabic document layout analysis. In Proceedings of the 2017 IEEE Eighth International Conference on Intelligent Computing and Information Systems (ICICIS 2017) (pp. 224-231). Cairo, Egypt: IEEE . DOI: 10.1109/INTELCIS.2017.8260051. © 2017 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.

Upload: others

Post on 12-Oct-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Deep Arabic Document Layout Analysis€¦ · Ibrahim M. Amer, Salma Hamdy and Mostafa G. M. Mostafa This is the accepted version of a conference paper published by IEEE. Amer, I.M.,

Deep Arabic Document Layout

Analysis

Ibrahim M. Amer, Salma Hamdy and Mostafa G. M.

Mostafa

This is the accepted version of a conference paper published by IEEE.

Amer, I.M., Hamdy, S. & Mostafa, M.G.M. (2017). Deep Arabic

document layout analysis. In Proceedings of the 2017 IEEE Eighth

International Conference on Intelligent Computing and Information

Systems (ICICIS 2017) (pp. 224-231). Cairo, Egypt: IEEE . DOI:

10.1109/INTELCIS.2017.8260051.

© 2017 IEEE. Personal use of this material is permitted. Permission from

IEEE must be obtained for all other uses, in any current or future media,

including reprinting/republishing this material for advertising or

promotional purposes, creating new collective works, for resale or

redistribution to servers or lists, or reuse of any copyrighted component

of this work in other works.

Page 2: Deep Arabic Document Layout Analysis€¦ · Ibrahim M. Amer, Salma Hamdy and Mostafa G. M. Mostafa This is the accepted version of a conference paper published by IEEE. Amer, I.M.,

Deep Arabic Document Layout Analysis1st Ibrahim M. Amer

Computer Science Department,Faculty of Computer

and Information Sciences,Ain Shams University,Cairo 11566, Egypt,

[email protected]

2nd Salma HamdyComputer Science Department,

Faculty of Computerand Information Sciences,

Ain Shams University,Cairo 11566, Egypt,

[email protected]

3rd Mostafa G. M. MostafaComputer Science Department,

Faculty of Computerand Information Sciences,

Ain Shams University,Cairo 11566, Egypt,

[email protected]

Abstract—Document layout analysis (DLA) is an essentialstep for Optical Character Recognition Systems (OCR). Thetext of the document fed to the OCR must be extracted firstand isolated from images if exist. The DLA task is difficult asthere is no fixed layout for all documents, but instead, thereare several layouts. There are various approaches for DLA forvarious different languages. In this paper, some of the previoustechniques used in this field will be listed and then we willdiscuss the proposed method that depends on deep learningfor documents’ text localization. We used Arabic Printed TextImage database (APTI [19]), ImageNet [18] and a datasetcollected from different Arabic newspapers for training andevaluation.

Keywords—Optical Character Recognition, Document Lay-out Analysis, Font Type Recognition, Font Size Recognition,Deep Learning, Deep Convolutional Neural Networks (D-CNN), Transfer Learning (TL).

I. INTRODUCTION

Document Layout Analysis is an important step for OCR,document management systems, document archiving systemsand more. OCR systems recognize printed or handwrittentext images, these images must contain text only and if thedocument contains text mixed with photos, graphs shapes orhalftones; this will result in a negative effect on recognitionaccuracy. Thus, DLA is a crucial step before OCR. DLAis a method of defining and recognizing the structure ofthe document or it is the way of categorizing the importantregions in the document. It includes separating its textcomponents from non-text to feed them directly to the OCR,identifying the title of the document if any, etc. Various fonttypes and sizes are also factors that can be used in documentanalysis, because if the font type/size provided to the OCRis different from what the OCR waits, the results might notbe precise. Reading order determination (right to left or leftto right) is also a necessary factor to enhance the DLA andproduce reliable results.

Document layout analysis has mainly two approaches. Thefirst is the bottom-up approach, which is concerned withparsing the document in an iterative way. Methods based on

bottom-up approach are concerned with grouping connectedregions into words then into text lines and text blocks. Thesecond approach is the top-down approach which is impliedfrom its name that the document is segmented into columnsfirst (relied on white space for example), segmenting its linesthen finally, segmenting its words. Several algorithms andtechniques are recently introduced for the DLA to documentswritten in English, but Arabic image documents did not getmuch attention compared with other languages.

This paper is organized as follows; section II summarizessome of the related works of Arabic document layout analysis.Section III shows the proposed technique. Finally, section IVdiscusses the achieved results and datasets used.

II. RELATED WORK

There are few papers that address the document analysis forArabic documents. This section discusses some of them.

Syed Saqib Bukhari et. al [1] proposed a technique for textlines extraction as well as determining reading order fromscanned documents written in various languages and differentstyles. The processing flow for their system has four mainsteps: Binarization, Text and Non-text segmentation, Text-lineDetection and Reading Order Determination. For binarizationstep, they used Otsu [2] and Sauvola [3] methods. The paperpresents an improvement for Bloomberg method in [6] fortext and non-text segmentation. Bloomberg’s method doesnot work well for non-text elements such as drawings. SyedSaqib Bukhari et.al. proposed two modifications in [7] to dealwith non-text elements. They used ridge-based line detectiontechnique for text line detection, which is robust againstnoise, skew; also, it can be used for documents with differentlanguages. Finally, the order of the text lines is detected anddetermined according to the reading flow (left to right or rightto left). They perform the reading order by applying someordering criteria according to Arabic reading order (right toleft).

Amany, Mohsen et. al in [8] proposed a method for Arabiclayout analysis. First, Sauvola binarization algorithm [3]is used for document binarization. Then, the images arecleaned up from noise induced by scanning process using

Page 3: Deep Arabic Document Layout Analysis€¦ · Ibrahim M. Amer, Salma Hamdy and Mostafa G. M. Mostafa This is the accepted version of a conference paper published by IEEE. Amer, I.M.,

median filters to reduce salt and pepper noise and Gaussiansmoothing filter to clear and eliminate both Gaussian andmarginal noise. After that, the document is deskewed usingRadon transform to correct its skew that might be induced bythe scanning process. Next, the document is segmented intozones and then each zone is classified into text or non-textusing Support Vector Machine (SVM) after extraction ofsome textual features. Finally, each zone is segmented intolines by clustering connected components into lines andwords via clustering the spaces between words and betweenstandalone characters of the same word using k-means. Thesystem has been tested and evaluated on Arabic historicaldocuments with different font types/sizes.

Syed Saqib Bukhari et.al in [9] presented a method forside-notes extraction in Arabic manuscripts. Feature extractionis relied mainly on connected component analysis. Everyconnected component is re-sized to size 64X64. For eachwindow, normalized height, foreground area, relative distanceand orientation is calculated to form a feature vector of size(4 + 64X64) including the window itself. The surroundingcontext area of the connected component is also included inthe feature vector. The final feature vector size is 8192. Finally,the side notes are classified using SVM. F-measure for main-body text segmentation is 95.02% and 94.68% for the sidenotes text.

III. PROPOSED METHOD

In this section, a method for Arabic DLA using deeplearning is proposed. Given a document image, the targetis to reveal, localize and isolate text regions from non-textones. A Deep Convolutional Neural Network (D-CNN) isused to classify document’s regions (zones) into text ornon-text. Given a document image, the document skew iscorrected first then the document is pre-processed usingAdaptive Run Length Smoothing Algorithm (ARLSA [10])to merge adjacent pixels together to form a block/zone.This zone can be classified later into text or non-text. Then,after applying the ARLSA algorithm, all the connectedcomponents are identified and localized using connectedcomponent analysis. After that, each block/zone is classifiedusing two techniques: 1) Zone Based Classification and2) Patch Based Classification, more specific details willbe mentioned in Classification subsection. Finally, each textzone is segmented into lines and words. The used dataset wascollected from Arabic newspapers (Al-Sharq [11], Riyadh[12] and Al-Youm [13]). Datasets will be mentioned in detailin the Results and Discussion section. In this section, eachcomponent of the proposed system will be discussed anddescribed in details. Proposed system architecture is shownin Fig. 1.

Input: DocumentImage

Pre-processing Skew CorrectionBinarization

ARLSA to form zonesConnected Components Analysis

Blocks/Zones Localization

Classificationready

document

Classification using CNN Zone Based ClassificationPatches Based Classification

Text/Non-text Zones

Segment Lines

Segment Words

Output: Lines &Words

Figure 1: Proposed System Architecture

1) Pre-processing:First, all documents are skew corrected using fast Houghtransform with block adjacency graph (BAG) [5] and binarizedusing Bradley’s adaptive thresholding method [4] using aneighborhood of size 1/8th of the image size. Then, ARLSAis applied to the binarized image. ARLSA is used to segmentthe document image into its building blocks (text and non-text regions) by grouping adjacent pixels together to form ablock/zone that can be fed to the classifier later in order toclassify this zone as text or non-text. ARLSA is better thanRLSA [14] because it can group in-homogeneous componentsand can work with documents written with variable font size.The effect of ARLSA or RLSA to the document image issimilar to dilating the image using a structuring element butwith applying some constraints as shown in Fig. 2. After that,bounding boxes of all segmented zones is identified usingconnected component analysis [15] [16] as in Fig. 3.

Page 4: Deep Arabic Document Layout Analysis€¦ · Ibrahim M. Amer, Salma Hamdy and Mostafa G. M. Mostafa This is the accepted version of a conference paper published by IEEE. Amer, I.M.,

Figure 2: Original document image (left) and the result afterARLSA (right).

Figure 3: Original document image (left) and the result afteridentifying connected components’ regions (right).

2) Classification: Since the dataset is manually collectedfrom newspapers available online in PDF form, labelingsuch a dataset to be used with a deep learning techniqueconsumes time and requires huge efforts, especially thatDL techniques, are categorized as data hungry techniques,

which require huge data for training. Transfer learning (TL)[17] concept is extremely helpful and useful in such cases.Transfer learning or knowledge transfer can be applied whenhaving a classification task and a domain with sufficienttraining data called the source domain DS , but there is noenough training data in another domain of interest calledthe target domain DT , where the latter domain may have adifferent feature space (FS). Self-driving cars for examplecan be trained using the approach of TL by preparing amodel that can segment and recognize cars, trucks andpedestrians in a computer simulated game; then, the modelcan be adapted to work with real images captured fromthe streets. In such cases, knowledge transfer if doneeffectively would incredibly enhance the performance oflearning by avoiding much expensive data labeling efforts.Given a classification task in one domain with a sufficienttraining data, the goal is to find the probability distributionof the target task with insufficient training samples usingthe information acquired from the domain with sufficient data.

The target is to classify regions (zones) into text or non-text regions using the dataset collected from the newspapersmentioned before. Since the used dataset set is small, we canuse transfer learning to enhance the accuracy and learning byusing a dataset in another domain but with sufficient trainingsamples. So, for the source domain DS , the famous ImageNet[18] dataset is used to extract from its images the features ofnon-textual components as all of its images are colorful imagesof nature scenes, animals, etc. For textual feature extraction,APTI dataset [19] is used to capture textual features from it.

So, having a source domain DS with a FS of XS capturedfrom ImageNet and APTI datasets with labels YS and a targetdomain DT with FS captured from a few samples collectedfrom the newspapers XT , now the objective is to find themarginal probability distribution of P (YT |XT ).

The whole training set for the source domain model DS

consists of 100,000 samples, 50,000 64X64 images from Im-ageNet and another 50,000 64X64 (after resize) images fromAPTI. 80,000 samples used for training and the remaining20,000 samples used for testing and validation. Mean has beensubtracted from all training samples to have a zero mean asin Equation 1.

X = (X − µ)/S (1)

Where X is the target feature, µ is the feature’s mean andS is the feature’s range of values m−n (in this case S = 255which is the maximum intensity value of the pixel).

A D-CNN network is used to train a classifier to classify textand non-text regions/zones. D-CNN has proven its ability forimage recognition tasks and have wide applications for videorecognition, NLP and speech recognition. The architecture ofthe network used is as follows: Two convolutional layers of64 feature maps of 3X3 kernel size each, followed by anotherconvolutional layer of 128 feature maps of 2X2 kernel sizeand a dropout = 0.5 [20] used to control overfitting. Finally,

Page 5: Deep Arabic Document Layout Analysis€¦ · Ibrahim M. Amer, Salma Hamdy and Mostafa G. M. Mostafa This is the accepted version of a conference paper published by IEEE. Amer, I.M.,

three consecutive fully connected layers of 1024, 512 and 512nodes respectively followed by a softmax layer of 2 nodes foroutput classes (0 for text and 1 for non-text).

Adam optimization algorithm [21] is used to update networkweights. The domain model DS achieved an accuracy =99.14%. TL in this case was applied because we have a smalland different dataset compared to the training set used in thesource domain model DS . After training the model on thesource domain DS , the model is re-trained on 10,000 trainingsamples that were collected from the Arabic newspapers(documents were sliced into patches of size 64X64) for thetarget model XT but after freezing all the layers of the pre-trained model DS . We froze all the layers to ensure thatwe do not unlearn the previously acquired knowledge fromthe pre-trained model. Results can be found at the end ofthis document (Fig. 8 and Fig. 9. Textual regions labeledin green and non-text regions labeled in red). As mentionedbefore, two techniques were used for classification, 1) zonebased classification and 2) patch based classification. Forzone/block based classification, ARLSA was used to segmentthe document image into zones/blocks then each block is slicedinto small patches of size 64X64. These patches are presentedto the network to classify each one of them as a text or non-text. After that, each zone/block is classified as text, if thenumber of the patches classified as text is greater than thehalf of the total number of patches of this zone as in Equation2.

Zonei =

{text, if m > 0.5× nnon-text, otherwise

(2)

Where m is the number of patches classified as text for eachzone, n is the total number of patches, i is the zone indexbecause the document may have more than one zone andthe parameter 0.5 was estimated by trial and error. For thistechnique, evaluation is based on number of zones classifiedas text or non-text.

For patches based classification, the same was done as thefirst technique by applying ARLSA first. Then each segmentedzone is sliced into small patches of size 64X64. The differenceis that these patches are fed to the CNN only without applyingany rules as the previous techniques. The evaluation of thistechnique is relied on the number of patches classified in thewhole 40 document images’ set.

3) Text lines and words segmentation: After labeling thedocument into text and non-text zones, text zones are seg-mented into lines and words. To segment lines, VerticalProjection Profile analysis (V PP ) was used. After text linessegmentation, each text line is segmented into words usingHorizontal Projection Profile analysis (HPP ) as in Equation3. V PP and HPP was used because the documents’ imagesthat we work on are clean from noise.

V PP (y) =∑

0<x<w

f(x, y);HPP (x) =∑

0<y<h

f(x, y) (3)

Where f(x, y) is the input image, w is the image width, his the image height, V PP (y) is the sum of each row of theimage and HPP (x) is the sum of each column of the image.Fig. 4 shows the V PP histogram. Lines can be segmentedat the lowest histogram responses (where response is equal tozero).

Figure 4: Vertical ProjectionProfile Histogram.

Figure 5: Horizontal ProjectionProfile Histogram.

Fig. 5 shows the HPP histogram. Words can be segmentedat the lowest histogram responses (when response is equal tozero).

IV. RESULTS AND DISCUSSION

A. DatasetsAs mentioned before, three datasets were used. ImageNet

[18], APTI [19] and a dataset collected from Arabic printednewspapers.

1) Arabic Printed Text Image database : it is a datasetfor Arabic printed words. It contains 113,284 word images ofseveral font types, styles and sizes. The database is used forcompetitions related to research of Arabic OCR. The databasecontains ten different font types with four font styles (Italic,Bold, Plain and Bold Italic) and with ten font sizes. APTIfonts are shown in Fig. 6.

Page 6: Deep Arabic Document Layout Analysis€¦ · Ibrahim M. Amer, Salma Hamdy and Mostafa G. M. Mostafa This is the accepted version of a conference paper published by IEEE. Amer, I.M.,

Figure 6: APTI database fonts: (A) Andalus, (B) Arabic Trans-parent, (C) AdvertisingBold, (D) Diwani Letter, (E) DecoTypeThuluth, (F) Simplified Arabic, (G) Tahoma, (H) TraditionalAatbic, (I) DecoType Naskh, (J) M Unicode Sara.

2) ImageNet: is a large visual dataset prepared for objectdetection and recognition. It contains more than 14,000,000images with 1000 different classes. The dataset has been usedfor years for their competitions and it became a very commondataset used for the tasks of image classification. The datasetwas used for the purpose of extraction of the visual features ofthe images regardless the category of the image itself. So, asthe target is to classify and differentiate between text and non-text components in the newspaper image, ImageNet was usedto train the D-CNN the structure of the images with naturalscenes, animals, etc. (non-text components).

3) Arabic Newspapers: A dataset of Arabic newspaperswas collected from three different Arabic newspapers’ web-sites (Al-Sharq [11], Riyadh [12] and Al-Youm [13]). Thecollected newspapers’ images were extracted from its PDFformat. All pages of the collected PDFs were converted intoimages to use them for training and testing in the proposedsystem. Dataset samples shown in Fig. 7.

Figure 7: From left to right: Al-Sharq, Al-Riyadh and Al-Youm newspapers used for training and testing.

B. Results

Training phase was done on a machine with AMD FX-8350 processor, 16 GB memory and NVidia GTX 1060 usedfor training the D-CNN. The training time of the D-CNN tookaround 28 continues hours. The system has been evaluated on40 document images (10 skewed and 30 non-skewed) collectedfrom Arabic newspapers. The evaluation was measured basedon two techniques as mentioned earlier: using 1) zone/blockbased classification and 2) patches based classification. Resultsfor both techniques in Table I and Table II. For zone basedclassification, total number of zones classified in 40 documents= 683; 220 zones are actual text zones and 463 zones are actualnon-text zones. For zone based classificatoin, total number ofzones classified in 40 documents = 683; 220 zones are actualtext zones and 463 zones are actual non-text zones.

Table I: Confusion matrix for zone based classification

Actualvalues

Prediction outcome

Text′ Non-text′ Total

Text 188 32 220

Non-text 98 365 463

Total 286 397

The achieved results for zone based classification:Precision = 0.66, Recall = 0.855 and F-score = 0.745.

For patches based classification, total number of patchesclassified in 40 documents = 8499; 5650 zones are actual textpatches and 2849 zones are actual non-text patches.

Table II: Confusion Matrix for Patches Based Classification

Actualvalues

Prediction outcome

Text′ Non-text′ Total

Text 4896 754 5650

Non-text 219 2630 2849

Total 5115 3384

Page 7: Deep Arabic Document Layout Analysis€¦ · Ibrahim M. Amer, Salma Hamdy and Mostafa G. M. Mostafa This is the accepted version of a conference paper published by IEEE. Amer, I.M.,

The achieved results for patches based classification:Precision = 0.96, Recall = 0.87 and F-score = 0.91.

V. CONCLUSION

We introduced a method for Arabic document layout analy-sis using deep learning. The main motivation of this research isto develop an algorithm for analyzing Arabic newspapers usingDeep Convolutional Neural Network (D-CNN) as it achievesgreat results for tasks related to machine vision. We showedthat using a proper pre-trained model and a small dataset wecould achieve good results by applying the concept of transferlearning.

We used two classification methods: 1) zone based classi-fication and 2) patches based classification. The system hasbeen evaluated on 40 document images (10 skewed and 30non-skewed); the best results achieved for text and non- textclassification are Precision = 0.96, Recall = 0.87 and F-score= 0.91. Using ImageNet and APTI datasets, we trained anetwork to discriminate images and text, and then we fine-tuned the network using a very small dataset collected fromArabic popular newspapers’ images. ImageNet was used toextract the visual features of the images regardless of theircategories because we only need to discriminate between textand non-text components; APTI dataset was used as well tocapture the visual features of the text.

REFERENCES

[1] Syed Saqib Bukhari et al., ”Layout Analysis of Arabic ScriptDocuments” in Guide to OCR for Arabic Scripts.London, UK:Springer, January 2012, ch 2, pp 35-53.

[2] Otsu, N.: A threshold selection method from gray-level histograms.IEEE Trans. Syst. Man Cybern. 9(1), 62?66 (1979.)

[3] Sauvola, Jaakko, and Matti Pietikinen. ”Adaptive document imagebinarization.” Pattern recognition 33, no. 2 (2000): 225-236.

[4] Bradley, Derek, and Gerhard Roth. ”Adaptive thresholding using theintegral image.” Journal of Graphics Tools 12, no. 2 (2007): 13-21.

[5] Singh, Chandan, Nitin Bhatia, and Amandeep Kaur. ”Hough transformbased fast skew detection and accurate skew correction methods.”Pattern Recognition 41, no. 12 (2008): 3528-3546.

[6] Bloomberg, D.S.: ”Multiresolution morphological approach to documentimage analysis”. In: Proceedings 1st International Conference onDocument Analysis and Recognition, St. Malo, France, pp. 963?971(1991.)

[7] Bukhari, S.S., Shafait, F., Breuel, T.M.: Improved document imagesegmentation algorithm using multiresolution morphology. In:Proceedings SPIE Document Recognition and Retrieval XVIII,San Jose, CA, USA, Jan. 2011.

[8] Hesham, Amany M., Mohsen AA Rashwan, Hassanin M. Al-Barhamtoshy, Sherif M. Abdou, Amr A. Badr, and Ibrahim Farag.”Arabic document layout analysis.” Pattern Analysis and Applications:1-13.

[9] Bukhari, Syed Saqib, Thomas M. Breuel, Abedelkadir Asi, and JihadEl-Sana. ”Layout analysis for arabic historical document images usingmachine learning.” In Frontiers in Handwriting Recognition (ICFHR),2012 International Conference on, pp. 639-644. IEEE, 2012.

[10] Nikolaou, Nikos, Michael Makridis, Basilis Gatos, NikolaosStamatopoulos, and Nikos Papamarkos. ”Segmentation of historicalmachine-printed documents using adaptive run length smoothing andskeleton segmentation paths.” Image and Vision Computing 28, no. 4(2010): 590-604.

[11] http://www.al-sharq.com. Last accessed February 2017

[12] http://www.alriyadh.com. Last accessed February 2017

[13] http://www.alyaum.com. Last accessed February 2017

[14] Wahl, Friedrich M., Kwan Y. Wong, and Richard G. Casey. ”Blocksegmentation and text extraction in mixed text/image documents.”Computer graphics and image processing 20, no. 4 (1982): 375-390.

[15] https://www.mathworks.com/help/images/ref/bwlabel.html. Lastaccessed July 2017

[16] https://www.mathworks.com/help/images/ref/regionprops.html. Lastaccessed July 2017

[17] Pan, Sinno Jialin, and Qiang Yang. ”A survey on transfer learning.”IEEE Transactions on knowledge and data engineering 22, no. 10(2010): 1345-1359.

[18] ] J. Deng, A. Berg, S. Satheesh, H. Su, A. Khosla, and L.FeiFei. ImageNet Large Scale Visual Recognition Competition 2012(ILSVRC2012). http://www.image-net.org/challenges/LSVRC/2012/last accessed July 2017.

[19] APTI Arabic Printed Text Image Database. https://diuf.unifr.ch/diva/APTI/index.html last accessed June 2017

[20] Srivastava, Nitish, Geoffrey E. Hinton, Alex Krizhevsky, Ilya Sutskever,and Ruslan Salakhutdinov. ”Dropout: a simple way to prevent neuralnetworks from overfitting.” Journal of Machine Learning Research 15,no. 1 (2014): 1929-1958.

[21] Kingma, Diederik, and Jimmy Ba. ”Adam: A method for stochasticoptimization.” arXiv preprint arXiv:1412.6980 (2014).

Page 8: Deep Arabic Document Layout Analysis€¦ · Ibrahim M. Amer, Salma Hamdy and Mostafa G. M. Mostafa This is the accepted version of a conference paper published by IEEE. Amer, I.M.,

(a) (b) (c)

(d) (e) (f)

Figure 8: (a), (d): original document images. (b), (e): zone based classification. (c), (f): patchesbased classification (green: text, red: non-text).

Page 9: Deep Arabic Document Layout Analysis€¦ · Ibrahim M. Amer, Salma Hamdy and Mostafa G. M. Mostafa This is the accepted version of a conference paper published by IEEE. Amer, I.M.,

a (b) c

d (e) f

Figure 9: (a), (d): original document images. (b), (e): zone based classification. (c), (f): patchesbased classification (green: text, red: non-text).