1,342 research outputs found

    Rotated Japanese Character Recognition

    Get PDF
    We proposed a rotated character recognition method using eigen-subspace for alpha-numeric characters so far. We first construct an eigen-subspace for each category using the covariance matrix calculated from a sufficient number of rotated character patterns. Next, we can obtain a locus by projecting their rotated characters onto the eigen sub-space and interpolating between their projected points. An unknown character is also projected onto the eigen subspace of each category. A single projection and multiple projections of the input character image were proposed. Then, the verification is carried out by calculating the distance between the projected points of the unknown character and the locus. Then the multiple projections showed a higher accuracy at low dimensions than a single projection for alphanumeric 62 categories. This time, we applied it for the first class of Japanese Industrial Standard (JIS) Kanji set which includes 2,965 categories. As the result, very high recognition accuracy over 99.8% was achieved by especially multiple projections of the input rotated images

    Kannada Character Recognition System A Review

    Full text link
    Intensive research has been done on optical character recognition ocr and a large number of articles have been published on this topic during the last few decades. Many commercial OCR systems are now available in the market, but most of these systems work for Roman, Chinese, Japanese and Arabic characters. There are no sufficient number of works on Indian language character recognition especially Kannada script among 12 major scripts in India. This paper presents a review of existing work on printed Kannada script and their results. The characteristics of Kannada script and Kannada Character Recognition System kcr are discussed in detail. Finally fusion at the classifier level is proposed to increase the recognition accuracy.Comment: 12 pages, 8 figure

    A Font Search Engine for Large Font Databases

    Get PDF
    A search engine for font recognition is presented and evaluated. The intended usage is the search in very large font databases. The input to the search engine is an image of a text line, and the output is the name of the font used when rendering the text. After pre-processing and segmentation of the input image, a local approach is used, where features are calculated for individual characters. The method is based on eigenimages calculated from edge filtered character images, which enables compact feature vectors that can be computed rapidly. In this study the database contains 2763 different fonts for the English alphabet. To resemble a real life situation, the proposed method is evaluated with printed and scanned text lines and character images. Our evaluation shows that for 99.1% of the queries, the correct font name can be found within the five best matches

    WordFences: Text localization and recognition

    Get PDF
    En col·laboració amb la Universitat de Barcelona (UB) i la Universitat Rovira i Virgili (URV)In recent years, text recognition has achieved remarkable success in recognizing scanned document text. However, word recognition in natural images is still an open problem, which generally requires time consuming post-processing steps. We present a novel architecture for individual word detection in scene images based on semantic segmentation. Our contributions are twofold: the concept of WordFence, which detects border areas surrounding each individual word and a unique pixelwise weighted softmax loss function which penalizes background and emphasizes small text regions. WordFence ensures that each word is detected individually, and the new loss function provides a strong training signal to both text and word border localization. The proposed technique avoids intensive post-processing by combining semantic word segmentation with a voting scheme for merging segmentations of multiple scales, producing an end-to-end word detection system. We achieve superior localization recall on common benchmark datasets - 92% recall on ICDAR11 and ICDAR13 and 63% recall on SVT. Furthermore, end-to-end word recognition achieves state-of-the-art 86% F-Score on ICDAR13

    Real-time gesture recognition using eigenvectors

    Get PDF
    This paper discusses an implementation for gesture recognition using eigenvectors under controlled conditions. This application of eigenvector recognition is trained on a set of defined hand images. Training images are processed using eigen techniques from the OpenCV image processing library. Test images are then compared in real-time. These techniques are outlined below.Telkom. CISCO, THRI

    Statistical Deformation Model for Handwritten Character Recognition

    Get PDF

    Scatteract: Automated extraction of data from scatter plots

    Full text link
    Charts are an excellent way to convey patterns and trends in data, but they do not facilitate further modeling of the data or close inspection of individual data points. We present a fully automated system for extracting the numerical values of data points from images of scatter plots. We use deep learning techniques to identify the key components of the chart, and optical character recognition together with robust regression to map from pixels to the coordinate system of the chart. We focus on scatter plots with linear scales, which already have several interesting challenges. Previous work has done fully automatic extraction for other types of charts, but to our knowledge this is the first approach that is fully automatic for scatter plots. Our method performs well, achieving successful data extraction on 89% of the plots in our test set.Comment: Submitted to ECML PKDD 2017 proceedings, 16 page
    corecore