10.07.2015 Views

Optical Character Recognition of Amharic Documents - CVIT - IIIT ...

Optical Character Recognition of Amharic Documents - CVIT - IIIT ...

Optical Character Recognition of Amharic Documents - CVIT - IIIT ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>Optical</strong> <strong>Character</strong> <strong>Recognition</strong> <strong>of</strong> <strong>Amharic</strong><strong>Documents</strong>Million Meshesha C. V. JawaharCenter for Visual Information TechnologyInternational Institute <strong>of</strong> Information TechnologyHyderabad - 500 032, Indiamillion@research.iiit.ac.in jawahar@iiit.ac.inAbstract— In Africa around 2,500 languagesare spoken. Some <strong>of</strong> these languages have theirown indigenous scripts. Accordingly, there is abulk <strong>of</strong> printed documents available in libraries,information centers, museums and <strong>of</strong>fices.Digitization <strong>of</strong> these documents enables toharness already available informationtechnologies to local information needs anddevelopments. This paper presents an <strong>Optical</strong><strong>Character</strong> <strong>Recognition</strong> (OCR) system forconverting digitized documents in locallanguages. An extensive literature surveyreveals that this is the first attempt that reportsthe challenges towards the recognition <strong>of</strong>indigenous African scripts and a possiblesolution for <strong>Amharic</strong> script. Research in therecognition <strong>of</strong> African indigenous scripts facesmajor challenges due to (i) the use <strong>of</strong> largenumber characters in the writing and (ii)existence <strong>of</strong> large set <strong>of</strong> visually similarcharacters. In this paper, we propose a novelfeature extraction scheme using principalcomponent and linear discriminant analysis,followed by a decision directed acyclic graphbased support vector machine classifier.<strong>Recognition</strong> results are presented on real-lifedegraded documents such as books, magazinesand newspapers to demonstrate theperformance <strong>of</strong> the recognizer.Index Terms— <strong>Optical</strong> <strong>Character</strong> <strong>Recognition</strong>;African Scripts; Feature Extraction; Classification;<strong>Amharic</strong> <strong>Documents</strong>.I. INTRODUCTIONNowadays, it is becoming increasinglyimportant to have information available indigital format for increased efficiency indata storage and retrieval, and opticalcharacter recognition (OCR) is being knownas one <strong>of</strong> valuable input devices in thisrespect [1], [2]. This is also supplementedby the emergency <strong>of</strong> large digital librariesand the advancement <strong>of</strong> informationtechnology.Digital libraries such as Universal DigitalLibrary, Digital Library <strong>of</strong> India, etc. areestablished for digitizing paper documentsand make them available to users via theInternet. The advancement <strong>of</strong> informationtechnology supports faster, more powerfulprocessors and high speed output devices(printers and others) to generate moreinformation at a faster rate. These days theavailability <strong>of</strong> relatively inexpensivedocument scanners and optical characterrecognition s<strong>of</strong>tware has made OCR anattractively priced data entry methodology[1].<strong>Optical</strong> character recognition technologywas invented in the early 1800s, when itwas patented as reading aids for the blind.In 1870, C. R. Carey patented an imagetransmission system using photocells, andin 1890 P.G. Nipkow invented sequentialscanning OCR. However, the practical OCRtechnology used for reading characters wasintroduced in the early 1950s as areplacement for keypunching system. Ayear later, D.H. Shephard developed thefirst commercial OCR for typewritten data.The 1980’s saw the emergence <strong>of</strong> OCRsystems intended for use with personal


computers [3]. Nowadays, it is common t<strong>of</strong>ind PC-based OCR systems that arecommercially available. However, most <strong>of</strong>these systems are developed to work withLatin-based scripts [4].<strong>Optical</strong> character recognition convertsscanned images <strong>of</strong> printed, typewritten orhandwritten documents into computerreadable format (such as ASCII, Unicode,etc.) so as to ease on-line data processing.The potential <strong>of</strong> OCR for data entryapplication is obvious: it <strong>of</strong>fers a faster,more automated, and presumably lessexpensive alternative to the manual dataentry devices, thereby improving theaccuracy and speed in transcribing datainto the computer system. Consequently, itincreases efficiency and effectiveness (byreducing cost, time and labor) in informationstorage and retrieval.Major applications <strong>of</strong> OCR include: (i)Library and <strong>of</strong>fice automation, (ii) Form andbank check processing, (iii) Documentreader systems for the visually impaired, (iv)Postal automation, and (v) Database andcorpus development for language modeling,text-mining and information retrieval [2], [5].While the use and application <strong>of</strong> OCRsystems is well developed for mostlanguages in the world that use both Latinand non-Latin scripts [6], an extensiveliterature survey reveals that fewconference papers are available on theindigenous scripts <strong>of</strong> African languages.Lately some research reports have beenpublished on <strong>Amharic</strong> OCR. <strong>Amharic</strong>character recognition is discussed in [7]with more emphasis to designing suitablefeature extraction scheme for scriptrepresentation. <strong>Recognition</strong> using directionfield tensor as a tool for <strong>Amharic</strong> charactersegmentation is also reported [8]. Workuand Fuchs [9] present handwritten <strong>Amharic</strong>bank check recognition. On the contrary,there are no similar works being found forother indigenous African scripts.Therefore, there is a need to exert mucheffort to come up with better and workableOCR technologies for African scripts inorder to satisfy the need for digitizedinformation processing in local languages.II. AMHARIC SCRIPTSIn Africa more than 2,500 languages,including regional dialects are spoken.Some are indigenous languages, whileothers are installed by conquerors <strong>of</strong> thepast. English, French, Portuguese, Spanishand Arabic are <strong>of</strong>ficial languages <strong>of</strong> many <strong>of</strong>the African countries. As a result, mostAfrican languages with a writing system usea modification <strong>of</strong> the Latin and Arabicscripts. There are also many languageswith their own indigenous scripts andwriting systems. Some <strong>of</strong> these scriptsinclude <strong>Amharic</strong> script (Ethiopia), Vai script(West Africa), Hieroglyphic script (Egypt),Bassa script (Liberia), Mende script (SierraLeone), Nsibidi/Nsibiri script (Nigeria andCameroon) and Meroitic script (Sudan) [10].<strong>Amharic</strong>, which belongs to the Semiticlanguage, became a dominant language inEthiopia back in history. It is the <strong>of</strong>ficial andworking language <strong>of</strong> Ethiopia and the mostcommonly learnt language next to Englishthroughout the country. Accordingly, thereis a bulk <strong>of</strong> information available in printedform that needs to be converted intoelectronic form for easy searching andretrieval as per users’ need. Suffice is tomention the huge amount <strong>of</strong> documentspiled high in information centers, libraries,museums and government and private


<strong>of</strong>fices in the form <strong>of</strong> correspondence letters,magazines, newspapers, pamphlets, books,etc. Converting these documents intoelectronic format is a must in order to (i)preserve historical documents, (ii) savestorage space, and (iii) enhance retrieval <strong>of</strong>relevant information via the Internet. Thisenables to harness existing informationtechnologies to local information needs anddevelopments.Fig. 1: <strong>Amharic</strong> alphabets (FIDEL) with their seven ordersrow-wise. The 2nd column shows list <strong>of</strong> basic charactersand others are vowels each <strong>of</strong> which derived from thebasic symbol.Those African languages using amodified version <strong>of</strong> Latin and Arabic scriptscan easily be integrated to the existing Latinand Arabic OCR technologies. It is worth tomention here some <strong>of</strong> the related worksreported at home and in the Diaspora topreserve African languages digitally(predominantly in languages that use Latinscripts). Corpora projects in Swahili (EastAfrica), open source s<strong>of</strong>tware for languageslike Zulu, Sepedi and Afrikaans (SouthAfrican), indigenous web browser inLuganda (Uganda), improvisation <strong>of</strong>keyboard characters for some Africanlanguages, the presence <strong>of</strong> Africanlanguages on the Internet, etc. [11].Therefore, we need to give moreemphasis to those indigenous Africanscripts. This is the motivation behind thepresent report. To the best <strong>of</strong> ourknowledge, this is the first work that reportsa robust <strong>Amharic</strong> character recognizer forconversion <strong>of</strong> printed document images <strong>of</strong>varying fonts, sizes, styles and quality.<strong>Amharic</strong> is written in the unique andancient Ethiopic script (inherited from Geez,a Semitic language), now effectively asyllabary requiring over 300 glyph shapesfor representation. As shown in Fig. 1,<strong>Amharic</strong> script has 33 core characters each<strong>of</strong> which occurs in seven orders: one basicform and six non-basic forms consisting <strong>of</strong>a consonant and following vowel. Vowelsare derived from basic characters withsome modification (like by attaching strokesat the middle or end <strong>of</strong> the base character,by elongating one <strong>of</strong> the leg <strong>of</strong> the basecharacter, etc.). Other symbols are alsoavailable that represent labialization,numerals, and punctuation marks. Thesebring the total number <strong>of</strong> characters in thescript to 310 as shown in Table 1.


D. Visual similarity <strong>of</strong> most characters inthe scriptThere are a number <strong>of</strong> very similarcharacters in <strong>Amharic</strong> script that are evendifficult for humans to identify them easily(examples are presented in Fig. 2). Robustdiscriminant features needs to be extractedfor classification <strong>of</strong> each <strong>of</strong> the characterinto their proper category or class.E. Language related issuesAfrican indigenous languages posemany additional challenges. Some <strong>of</strong> theseare: (i) lack <strong>of</strong> standard representation forthe fonts and encoding, (ii) lack <strong>of</strong> supportfrom operating systems, browsers andkeyboard, and (iii) lack <strong>of</strong> languageprocessing routines.These issues add to the complexity <strong>of</strong>the design and implementation <strong>of</strong> an opticalcharacter recognition system.SegmentationComponentsFeatureExtractionFeatureVectorsClassifierRefinesImagesModelPreprocessingRawImagesModelModelRecognizerConvertedTextDocumentImagesDocumentTextCorrectedTextPost-processingFig. 3: Overview <strong>of</strong> the <strong>Amharic</strong> OCR design.III. RECOGNITION OF AMHARIC SCRIPTWe built an <strong>Amharic</strong> OCR on top <strong>of</strong> anearlier work for Indian languages [12] andalso reported a preliminary result <strong>of</strong> ourwork in [13]. Fig. 3 shows the generalframework <strong>of</strong> <strong>Amharic</strong> OCR design.The system accepts as input scanneddocument images. The resolution, ornumber <strong>of</strong> dots per inch (dpi) produced bythe scanner determines how closely thereproduced image corresponds to theoriginal. The higher the resolution, the moreinformation is recorded and, therefore, thegreater the file size and vice versa. Theresolution most suitable for characterrecognition ranges from 200 dpi to 600 dpidepending on document quality. Actuallymost systems recommend to scandocuments at a resolution <strong>of</strong> 300 dpi foroptimum OCR accuracy. We also use thisresolution to digitize documents for theexperiment.Once images are captured, they arepreprocessed before individual components


different representations <strong>of</strong> the characters.Some <strong>of</strong> these features consider pr<strong>of</strong>iles,structural descriptors and transform domainrepresentations [2], [16]. Alternates onecould consider the entire image as thefeature. The former methods are highlylanguage specific and become verycomplex to represent all the characters inthe script. The later scheme providesexcellent results for printed characterrecognition.As a result, we extract features from theentire image by concatenating all the rowsto form a single contiguous vector. Thisfeature vector consists <strong>of</strong> zeros (0s) andones (1s) representing background andforeground pixels in the image, respectively.With such a representation, memory andcomputational requirements are veryintensive for languages like <strong>Amharic</strong> thathave large number <strong>of</strong> characters in thewriting.Therefore we need to transform thefeatures to obtain a lower dimensionalrepresentation. There are various methodsemployed in pattern recognition literaturesfor reducing the dimension <strong>of</strong> featurevectors [17]. We consider in the presentwork Principal Component Analysis (PCA)and Linear Discriminant Analysis (LDA).4.1 Principal Component AnalysisPrincipal component analysis (PCA) canhelp identify new features (in a lowerdimensional subspace) that are most usefulfor representation [17]. This should be donewithout losing valuable information.Principal components can give superiorperformance for font-independent OCRs,easy adaptation across languages, andscope for extension to handwrittendocuments.Consider thethi image samplerepresented as an M dimensional (column)vector xi, where M depends on the imagesize. From the large sets <strong>of</strong> trainingdatasets, x , , x Nwe compute thecovariance matrix as:1∑ =NN∑i 11 L T( xi− µ )( xi− µ )= (1)Then we need to identify minimal dimension,say K such that:K∑ − 1λ ∑i i/ Trace()thi≤ α (2)where λiis the largest eigenvalue <strong>of</strong> thecovariance matrix ∑ and α is a limitingvalue in percent.Eigenvectors corresponding to thelargest K eigenvalues are the direction <strong>of</strong>thgreatest variance. The k eigenvector isthe direction <strong>of</strong> greatest variationstperpendicular to the first through ( k −1)eigenvectors. The eigenvectors arearranged as rows in matrix A and thistransformation matrix is used to computethe new feature vector by projecting asY = . With this we get the best oneiAx idimensional representation <strong>of</strong> thecomponent images with reduced featuresize.Principal component analysis (PCA)yields projection directions that maximizethe total scatter across all classes. Inchoosing the projection which maximizestotal scatter, PCA retains not only betweenclassscatter that is useful for classification,but also within-class scatter that isunwanted information for classificationpurposes. Much <strong>of</strong> the variation seenamong document images is due to printingvariations and degradations. If PCA isapplied on such images, the transformation


matrix will contain principal componentsthat retain these variations in the projectedfeature space. Consequently, the points inthe projected space will not be wellseparated and the classes may be smearedtogether.Thus, while the PCA projections areoptimal for representation in a lowdimensional space, they may not be optimalfrom a discrimination standpoint forclassification. Hence, in our work, wefurther apply linear discriminant analysis toextract useful features from the reduceddimensional space using PCA.4.2. Linear Discriminant AnalysisLinear discriminant analysis (LDA)selects features based entirely upon theirdiscriminatory potential. Let the projectionbe y = Dx , where x is the input image andD is the transformation matrix. Theobjective <strong>of</strong> LDA is to find a projection thatmaximizes the ratio <strong>of</strong> the between-classscatter and the within-class scatter [15].We apply the algorithm proposed byFoley and Sammon [18] for lineardiscriminant analysis. The algorithmextracts a set <strong>of</strong> optimal discriminantfeatures for a two-class problem which suitsthe Support Vector Machine (SVM)classifier we used for classification task.thConsider the i image samplerepresented as an M dimensional (column)vector xi, where M is the reduceddimension using PCA. From the given sets<strong>of</strong> training components, x , L 1, x Nwecompute with-in class scatter ( W ) matrixand between-class difference ( ∆ ) as:∆ = µ1− µ 2ththwhere is the j sample in i class.x ij(4)Sum <strong>of</strong> the with-in class scatter is alsodetermined by,A = cW1+ 1−c)( W2(5)where 0 ≤ c ≤1, and the scatter space using:1−11Sij= d−1d = α A ∆Sfor1= 1/ SdSniAn = 2n−1n11−1= α Ann−1dtot 2ω = ( d ∆)/ d Ad1 ⎡cnS= ⎢cn⎢⎣jK−1'{ ∆ − [ d Ld] S [ 1/ α 0L0]}tn−1n−11n+ S−1n−1n−1n n−1t− y Sn−1tnn−1y y S−1n−11−1− S ⎤n−1yn⎥1 ⎥⎦(6)Fig. 4: Algorithm for computing L best discriminantfeature vectors. The vertical and horizontal lines are drawn−1to partition the scatter matrix S n such that the resultobtained is a2× 2 matrix.The algorithm presented in Fig. 4 iscalled recursively for extracting an optimalset <strong>of</strong> discriminant vectors ( d n) thatcorresponds to the first L highestdiscriminant values such that( ω1 ≥ ω2≥ L ≥ ωL≥ 0 ) [18]. Here, K is thenumber <strong>of</strong> iterations for computingdiscriminant vectors as well as discriminantvalues, and, α is chosen such that:nWiNit( xi ij− µ )( xi ij− µi)1= ∑ =(3)d td n n=1(7)


ynandmanner.ycnnc n= L= sare determined in the following[ SinS( n−1)(n)nn− ytnS−1n−1yn](8)Since the principal purpose <strong>of</strong>classification is discrimination, thediscriminant vector subspace <strong>of</strong>fersconsiderable potential as feature extractiontransformation. Hence, the use <strong>of</strong> lineardiscriminant features for each pair <strong>of</strong>classes in each component classification isan attractive solution to design betterclassifiers.However, the problem with thisapproach is that the classification becomesslower, because LDA extracts features foreach pair-wise class. This requires stackspace for storing N ( N −1)/ 2 transformationmatrices. For instance, for dataset with 200classes LDA generates around 20 thousandtransformation matrices for each pair and isvery expensive for large class problems.Hence, we propose a two-stage featureextraction scheme; PCA followed by LDAfor optimal discriminant feature extraction.We first reduce the feature dimension usingPCA and then run LDA on the reducedlower dimensional space to extract the mostdiscriminant feature vector. This reduces toa great extent the storage andcomputational complexity, while enhancingthe performance <strong>of</strong> SVM-based decisiondirected acyclic graph (DDAG) classifier.V. CLASSIFICATIONTraining and testing are the two basicphases <strong>of</strong> any pattern classification problem.During training phase, the classifier learnsthe association between samples and theirlabels from labeled samples. The testingphase involves analysis <strong>of</strong> errors in theclassification <strong>of</strong> unlabelled samples in orderto evaluate classifier’s performance. Ingeneral it is desirable to have a classifierwith minimal test error.We use Support Vector Machine (SVM)for classification task [19]. SVMs are pairwisediscriminating classifiers with theability to identify the decision boundary withmaximal margin. Maximal margin results inbetter generalization [20] which is a highlydesirable property for a classifier to performwell on a novel data set. Support vectormachines are less complex and performbetter (lower actual error) with limitedtraining data.Identification <strong>of</strong> the optimal hyper-planefor separation involves maximization <strong>of</strong> anappropriate objective function. The result <strong>of</strong>the training phase is the identification <strong>of</strong> aset <strong>of</strong> labeled support vectors and a set<strong>of</strong> coefficients αi.Support vectors are thesamples near the decision boundary. Theyhave class labels y iranging ±1. Thedecision is made from:l∑f ( x)= sgn( α y K(x , x))i=1iiix i(9)where K is the kernel function, which isdefined by:K( x,y)= φ(x)φ(y)(10)where φ : R d → H maps the data points inlower dimensions to a higher dimensionalspace H.Binary classifiers like SVM are basicallydesigned for two class classificationproblems. However, because <strong>of</strong> theexistence <strong>of</strong> a number <strong>of</strong> characters in anyscript, optical character recognition problemis inherently multi-class in nature. The field<strong>of</strong> binary classification is mature, and


provides a variety <strong>of</strong> approaches to solvethe problem <strong>of</strong> multi-class classification [21].Most <strong>of</strong> the existing multi-class algorithmsaddress the problem by first dividing it intosmaller sets <strong>of</strong> a pair <strong>of</strong> classes and thencombine the results <strong>of</strong> binary classifiersusing a suitable voting methods such asmajority or weighted majority approaches[21], [22].Fig. 5: A rooted binary decision directed acyclic graph (DDAG) multi-class classifier for eight-class classificationproblem.5.1 Multi-class SVMsMulti-class SVMs are usuallyimplemented as combinations <strong>of</strong> two-classsolution using majority voting methods. Ithas been shown that the integration <strong>of</strong>pairwise classifiers using decision directedacyclic graph (DDAG) results in betterperformance as compared to other populartechniques such as decision tree, Bayesian,etc. [19].We construct a decision directed acyclicgraph (DDAG), where each node in thegraph corresponds to a two-class classifierfor a pair <strong>of</strong> classes. The multi-classclassifier built using decision directedacyclic graph (DDAG) for 8-classclassification problem is shown in Fig. 5. Itcan be observed that the number <strong>of</strong> binaryclassifiers built for a N class classificationproblem is N ( N −1) / 2.The input vector ispresented at the root node <strong>of</strong> the DDAGand moves through the DDAG until itreaches the leaf node where the output(class label) is obtained. At each node adecision is made concerning to which classthe input belongs.For example, at the root node <strong>of</strong> Fig. 5,a decision is made whether the inputbelongs to class 1 or class 8. If it doesnotbelong to class 1, it moves to the left child;otherwise, if it doesnot belong to class 8, itmoves to the right child. This process isrepeated at each <strong>of</strong> the remaining nodes tillit reaches the leaf node where the decisionis reached to classify the input as one <strong>of</strong> theeight classes.


VI. RESULTS AND DISCUSSIONSWe have a recognition system thatconverts a given document images intoequivalent textual format. The systemaccepts either already scanned documentimages or scan a given <strong>Amharic</strong> textdocument at a resolution <strong>of</strong> 300 dpi on aflat-bed scanner, HP7670 Scanjet scanner.The scanned document is binarized, noiseremoved, skew corrected and scaled beforeindividual components are extracted.Preprocessed pages are then segmentedinto character components for featureextraction and classification. Following<strong>Amharic</strong> letters shape formation; we canfirst decompose lines in a text page intowords and then words into appropriatecharacter components for recognition andthen recompose the recognizedcomponents to determine the order <strong>of</strong>characters, words and lines in a text page.Once character components areidentified, optimal discriminant features (ina lower dimensionality space) are extractedfor classification. We use a two stagedimensionality reduction scheme based on99 percent principal component analysiswhich is followed by 15 percent lineardiscriminant analysis.The original dimension <strong>of</strong> feature vector<strong>of</strong> each character image (afternormalization) is 400 ( 20× 20 ). Principalcomponent analysis reduces thedimensionality from 400 to 295 and lineardiscriminant analysis further to 50. We aredealing with such reduced optimaldiscriminant feature vectors. Both methodsperform well in feature dimensionalityreductions. The use <strong>of</strong> a two-stage featureextraction scheme further solves thesimilarity problem encountered during theapplication <strong>of</strong> PCA alone. This is because<strong>of</strong> the fact that the new scheme extractsoptimal features that discriminate betweena pair <strong>of</strong> characters employed forclassification using support vector machine.We conduct extensive experiments toevaluate the performance <strong>of</strong> the recognitionprocess on the various datasets <strong>of</strong> <strong>Amharic</strong>scripts. The experiments are organized in asystematic manner considering the varioussituations encountered in real-life printeddocuments. Our datasets are <strong>of</strong> two types:One set <strong>of</strong> datasets considers printingvariations (such as fonts, styles and sizes).The other is degraded documents such asnewspapers, magazines and books. Wereport the performance <strong>of</strong> the recognizer inall these datasets and the result obtained ispromising to extend it for other indigenousAfrican scripts.ResultPowerGeezFonts Point Size StyleVisualAgafari Alpas 10 12 14 16 Normal Bold ItalicGeezDatasets 7850 7850 7850 7850 7680 7680 7680 7680 7680 7680 7680Accuracy 99.08 96.24 95.53 95.16 98.64 99.08 98.06 98.21 99.08 98.21 89.67Table 2: Performance result <strong>of</strong> the <strong>Amharic</strong> OCR on pages that varies in fonts, sizes and styles.In the first experiment, we consider theprinting variation. We test on the mostpopular fonts such as PowerGeez,VisualGeez, Agafari and Alpas that areused for typing and printing purposes, fourpoint sizes (10, 12, 14 and 16) and fontstyles (such as normal, bold and italics).Performance results are shown in Table 2.


[6] Suen, C. Y., Mori, S., Kim, S. H. and Leung, C.H, “Analysis and recognition <strong>of</strong> Asian scripts -The state <strong>of</strong> the art”, in International Conferenceon Document Analysis and <strong>Recognition</strong>, 2005,pp. 866-878.[7] Cowell, J. and Hussain, F., “<strong>Amharic</strong> <strong>Character</strong><strong>Recognition</strong> using a Fast Signature BasedAlgorithm,” in Proceedings <strong>of</strong> the SeventhInternational Conference on InformationVisualization, 2003, pp. 384-389.[8] Yaregal, A. and Bigun, J., “Ethiopic <strong>Character</strong><strong>Recognition</strong> Using Direction Field Tensor,“ in18th International Conference on Pattern<strong>Recognition</strong> (ICPR), 2006, pp. 284-287.[9] Worku, A., and Fuchs, S., “Handwritten <strong>Amharic</strong>bank check recognition using hidden markovrandom field”, in Document Image Analysis andRetrieval Workshop, 2003, p. 28.[10] Mafundikwa, S., “African alphabets,”www.ziva.org.zw, Link to a page, November2000 [Online]. Available:http://www.ziva.org.zw/afrikan.htm, [AccessedApril 23, 2007].[11] Osborn, D., "African languages and informationand communication technologies: literacy,access, and the future". in Proceedings <strong>of</strong> the35 th Annual Conference on African Linguistics,2006, pp. 86 - 93.[12] Jawahar, C. V., Pavan Kumar, M. N. S. S. K.and Ravi Kiran, S. S., “A bilingual OCR for Hindi-Telugu documents and its applications”, inInternational Conference on Document Analysisand <strong>Recognition</strong> (ICDAR), 2003, pp. 408-412.[13] Million Meshesha and Jawahar, C. V.,“<strong>Recognition</strong> <strong>of</strong> printed <strong>Amharic</strong> documents,” inProceedings <strong>of</strong> Eighth International Conferenceon Document Analysis and <strong>Recognition</strong> (ICDAR),2005, pp 784-788.[14] Casey, R. G. and Lecoline, E., “A survey <strong>of</strong>methods and strategies in charactersegmentation,” IEEE Transactions on PatternAnalysis and Machine Intelligence, vol. 18, no. 7,July 1996, pp. 690-706.[15] Trier, O., Jain, A. and Taxt, T., "Featureextraction methods for character recognition - Asurvey", Pattern <strong>Recognition</strong>, vol 29, no. 4, 1996,pp. 641-662.[16] Gonzalez, R. C. and Woods, R. E., Digital ImageProcessing, 2nd ed., India: Pearson Education,2002.[17] .Duda, R. O., Hart, P. E. and Stork, D. G.,Pattern Classification, New York: John Wiley &Sons, Inc., 2001.[18] Foley, D.H. and Sammon, J. W., “An optimal set<strong>of</strong> discriminant vectors”, IEEE Transactions onComputing, vol. 24, March 1975, pp. 271-278.[19] Platt, J. C., Cristianini, N. and Shawe-Taylor, J.,“Large margin DAGs for multi-classclassification”, in Advances in Neural InformationProcessing Systems 12, 2000, pp. 547-553.[20] Burges, C. JC., “A tutorial on support vectormachines for pattern recognition,” Data Miningand Knowledge Discovery, vol. 2, no. 2, June1998, pp. 121-167.[21] Allwein, E. L., Schapire, R. E. and Singer, Y.,“Reducing multiclass to binary: A unifyingapproach for margin classification,” InProceeding <strong>of</strong> 17th International Conf. onMachine Learning, 2000, pp. 9-16.[22] Dietterich, T. G., “Ensemble methods in machinelearning,” Lecture Notes in Computer Science,vol. 1857, 2000, pp. 1-15.Million Meshesha received his M. Sc. from AddisAbaba University (AAU), Ethiopia, in 2000. Since then,he is with the Faculty <strong>of</strong> Informatics at AAU. Presently,he is a PhD candidate at the International Institute <strong>of</strong>Information Technology (<strong>IIIT</strong>) Hyderabad, India. Hisresearch interests include Document Image Analysis,Artificial Intelligence and Information Retrieval.C. V. Jawahar received his PhD in 1998 from IITKharagpur, India. He worked with Center for ArtificialIntelligence and Robotics till Dec 2000 and since thenhe is with <strong>IIIT</strong> Hyderabad, India. He is presently anAssociate Pr<strong>of</strong>essor. His areas <strong>of</strong> research areDocument Image Analysis, Content Based InformationRetrieval and Computer Vision.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!