Show simple item record

dc.contributor.advisorDavis, Larryen_US
dc.contributor.advisorDoermann, Daviden_US
dc.contributor.authorAgrawal, Muditen_US
dc.date.accessioned2011-10-08T05:34:18Z
dc.date.available2011-10-08T05:34:18Z
dc.date.issued2011en_US
dc.identifier.urihttp://hdl.handle.net/1903/11875
dc.description.abstractLarge scale document digitization projects continue to motivate interesting document understanding technologies such as script and language identification, page classification, segmentation and enhancement. Typically, however, solutions are still limited to narrow domains or regular formats such as books, forms, articles or letters and operate best on clean documents scanned in a controlled environment. More general collections of heterogeneous documents challenge the basic assumptions of state-of-the-art technology regarding quality, script, content and layout. Our work explores the use of adaptive algorithms for the automated analysis of noisy and complex document collections. We first propose, implement and evaluate an adaptive clutter detection and removal technique for complex binary documents. Our distance transform based technique aims to remove irregular and independent unwanted foreground content while leaving text content untouched. The novelty of this approach is in its determination of best approximation to clutter-content boundary with text like structures. Second, we describe a page segmentation technique called Voronoi++ for complex layouts which builds upon the state-of-the-art method proposed by Kise [Kise1999]. Our approach does not assume structured text zones and is designed to handle multi-lingual text in both handwritten and printed form. Voronoi++ is a dynamically adaptive and contextually aware approach that considers components' separation features combined with Docstrum [O'Gorman1993] based angular and neighborhood features to form provisional zone hypotheses. These provisional zones are then verified based on the context built from local separation and high-level content features. Finally, our research proposes a generic model to segment and to recognize characters for any complex syllabic or non-syllabic script, using font-models. This concept is based on the fact that font files contain all the information necessary to render text and thus a model for how to decompose them. Instead of script-specific routines, this work is a step towards a generic character and recognition scheme for both Latin and non-Latin scripts.en_US
dc.titleAdaptive Algorithms for Automated Processing of Document Imagesen_US
dc.typeDissertationen_US
dc.contributor.publisherDigital Repository at the University of Marylanden_US
dc.contributor.publisherUniversity of Maryland (College Park, Md.)en_US
dc.contributor.departmentComputer Scienceen_US
dc.subject.pqcontrolledComputer scienceen_US
dc.subject.pquncontrolleddocument image processingen_US
dc.subject.pquncontrolledmachine learningen_US
dc.subject.pquncontrollednoise removalen_US
dc.subject.pquncontrolledoptical character recognitionen_US
dc.subject.pquncontrolledpage segmentationen_US
dc.subject.pquncontrolledpattern recognitionen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record