15 research outputs found

    Metrics for Complete Evaluation of OCR Performance

    Get PDF
    International audienceIn this paper, we study metrics for evaluating OCR performance both in terms of physical segmentation and in terms of textual content recognition. These metrics rely on the OCR output (hypothesis) and the reference (also called ground truth) input format. Two evaluation criteria are considered: the quality of segmentation and the character recognition rate. Three pairs of input formats are selected among two types of inputs: text only (text) and text with spatial information (xml). These pairs of inputs reference-to-hypothesis are: 1) text-to-text, 2) xml-to-xml and 3) text-to-xml. For the text-to-text pair, we selected the RETAS method to perform experiments and show its limits. Regarding text-to-xml, a new method based on unique word anchors is proposed to solve the problem of aligning texts with different information. We define the ZoneMapAltCnt metric for the xml-to-xml approach and show that it offers the most reliable and complete evaluation compared to the other two. Open source OCRs like Tesseract and OCRopus are selected to perform experiments. The datasets used are a collection of documents from the ISTEX 1 document database, from French newspaper "Le Nouvel Observateur" as well as invoices and administrative document gathered from different collaborations

    Optical Character Recognition Using Morphological Attributes.

    Get PDF
    This dissertation addresses a fundamental computational strategy in image processing hand written English characters using traditional parallel computers. Image acquisition and processing is becoming a thriving industry because of the frequent availability of fax machines, video digitizers, flat-bed scanners, hand scanners, color scanners, and other image input devices that are now accessible to everyone. Optical Character Recognition (OCR) research increased as the technology for a robust OCR system became realistic. There is no commercial effective recognition system that is able to translate raw digital images of hand written text into pure ASCII. The reason is that a digital image comprises of a vast number of pixels. The traditional approach of processing the huge collection of pixel information is quite slow and cumbersome. In this dissertation we developed an approach and theory for a fast robust OCR system for images of hand written characters using morphological attribute features that are expected by the alphabet character set. By extracting specific morphological attributes from the scanned image, the dynamic OCR system is able to generalize and approximate similar images. This generalization is achieved with the usage of fuzzy logic and neural network. Since the main requirement for a commercially effective OCR is a fast and a high recognition rate system, the approach taken in this research is to shift the recognition computation into the system\u27s architecture and its learning phase. The recognition process constituted mainly simple integer computation, a preferred computation on digital computers. In essence, the system maintains the attribute envelope boundary upon which each English character could fall under. This boundary is based on extreme attributes extracted from images introduced to the system beforehand. The theory was implemented both on a SIMD-MC\sp2 and a SISD machine. The resultant system proved to be a fast robust dynamic system, given that a suitable learning had taken place. The principle contributions of this dissertation are: (1) Improving existing thinning algorithms for image preprocessing. (2) Development of an on-line cluster partitioning procedure for region oriented segmentation. (3) Expansion of a fuzzy knowledge base theory to maintain morphological attributes on digital computers. (4) Dynamic Fuzzy learning/recognition technique

    Crowdsourcing in cultural heritage

    Get PDF
    The aims of this study, within the framework of the Europeana Common Culture project are to: 1. Determine current and planned approaches and practices within the Europeana aggregation ecosystem in relation to crowdsourced metadata and content. 2. Investigate, as comprehensively as possible, past and existing DCH crowdsourcing initiatives across Europe, systematically describing their status and gaining a sound understanding of current practices. 3. Assess the feasibility, desirability and challenges faced in any effort to strengthen the pipeline from such initiatives to enable ingestion of their metadata or access to their content through Europeana. 4. Provide recommendations and guidelines for consideration by Europeana, aggregators and Cultural Heritage Institutions. 5. Support the creation of training materials for the Europeana ecosystem in terms of any agreed interaction with Europeana around crowdsourced assets and deliver this by suitable means (e.g. webinars, Europeana Pro). The work carried out has involved a 9 month programme (April-December 2020) consisting of desk research, , three online questionnaire surveys (to national aggregators; thematic/domain aggregators and external crowdsourcing initiatives respectively), a series of interviews and three consultative on-line events. The survey data are summarised in extensive annexes

    Policy-Gradient Algorithms for Partially Observable Markov Decision Processes

    No full text
    Partially observable Markov decision processes are interesting because of their ability to model most conceivable real-world learning problems, for example, robot navigation, driving a car, speech recognition, stock trading, and playing games. The downside of this generality is that exact algorithms are computationally intractable. Such computational complexity motivates approximate approaches. One such class of algorithms are the so-called policy-gradient methods from reinforcement learning. They seek to adjust the parameters of an agent in the direction that maximises the long-term average of a reward signal. Policy-gradient methods are attractive as a \emph{scalable} approach for controlling partially observable Markov decision processes (POMDPs). In the most general case POMDP policies require some form of internal state, or memory, in order to act optimally. Policy-gradient methods have shown promise for problems admitting memory-less policies but have been less successful when memory is required. This thesis develops several improved algorithms for learning policies with memory in an infinite-horizon setting. Directly, when the dynamics of the world are known, and via Monte-Carlo methods otherwise. The algorithms simultaneously learn how to act and what to remember. ..

    Policy-Gradient Algorithms for Partially Observable Markov Decision Processes

    No full text
    Partially observable Markov decision processes are interesting because of their ability to model most conceivable real-world learning problems, for example, robot navigation, driving a car, speech recognition, stock trading, and playing games. The downside of this generality is that exact algorithms are computationally intractable. Such computational complexity motivates approximate approaches. One such class of algorithms are the so-called policy-gradient methods from reinforcement learning. They seek to adjust the parameters of an agent in the direction that maximises the long-term average of a reward signal. Policy-gradient methods are attractive as a \emph{scalable} approach for controlling partially observable Markov decision processes (POMDPs). In the most general case POMDP policies require some form of internal state, or memory, in order to act optimally. Policy-gradient methods have shown promise for problems admitting memory-less policies but have been less successful when memory is required. This thesis develops several improved algorithms for learning policies with memory in an infinite-horizon setting. Directly, when the dynamics of the world are known, and via Monte-Carlo methods otherwise. The algorithms simultaneously learn how to act and what to remember. ..
    corecore