1,342 research outputs found
Rotated Japanese Character Recognition
We proposed a rotated character recognition
method using eigen-subspace for alpha-numeric characters so
far. We first construct an eigen-subspace for each category
using the covariance matrix calculated from a sufficient
number of rotated character patterns. Next, we can obtain a
locus by projecting their rotated characters onto the eigen sub-space
and interpolating between their projected points. An
unknown character is also projected onto the eigen subspace
of each category. A single projection and multiple projections
of the input character image were proposed. Then, the
verification is carried out by calculating the distance between
the projected points of the unknown character and the locus.
Then the multiple projections showed a higher accuracy at low
dimensions than a single projection for alphanumeric 62
categories. This time, we applied it for the first class of
Japanese Industrial Standard (JIS) Kanji set which includes
2,965 categories. As the result, very high recognition accuracy
over 99.8% was achieved by especially multiple projections of
the input rotated images
Kannada Character Recognition System A Review
Intensive research has been done on optical character recognition ocr and a
large number of articles have been published on this topic during the last few
decades. Many commercial OCR systems are now available in the market, but most
of these systems work for Roman, Chinese, Japanese and Arabic characters. There
are no sufficient number of works on Indian language character recognition
especially Kannada script among 12 major scripts in India. This paper presents
a review of existing work on printed Kannada script and their results. The
characteristics of Kannada script and Kannada Character Recognition System kcr
are discussed in detail. Finally fusion at the classifier level is proposed to
increase the recognition accuracy.Comment: 12 pages, 8 figure
A Font Search Engine for Large Font Databases
A search engine for font recognition is presented and evaluated. The intended usage is the search in very large font databases. The input to the search engine is an image of a text line, and the output is the name of the font used when rendering the text. After pre-processing and segmentation of the input image, a local approach is used, where features are calculated for individual characters. The method is based on eigenimages calculated from edge filtered character images, which enables compact feature vectors that can be computed rapidly. In this study the database contains 2763 different fonts for the English alphabet. To resemble a real life situation, the proposed method is evaluated with printed and scanned text lines and character images. Our evaluation shows that for 99.1% of the queries, the correct font name can be found within the five best matches
WordFences: Text localization and recognition
En col·laboració amb la Universitat de Barcelona (UB) i la Universitat Rovira i Virgili (URV)In recent years, text recognition has achieved remarkable success in recognizing scanned
document text. However, word recognition in natural images is still an open problem,
which generally requires time consuming post-processing steps. We present a novel architecture
for individual word detection in scene images based on semantic segmentation.
Our contributions are twofold: the concept of WordFence, which detects border areas
surrounding each individual word and a unique pixelwise weighted softmax loss function
which penalizes background and emphasizes small text regions. WordFence ensures that
each word is detected individually, and the new loss function provides a strong training
signal to both text and word border localization. The proposed technique avoids intensive
post-processing by combining semantic word segmentation with a voting scheme
for merging segmentations of multiple scales, producing an end-to-end word detection
system. We achieve superior localization recall on common benchmark datasets - 92%
recall on ICDAR11 and ICDAR13 and 63% recall on SVT. Furthermore, end-to-end
word recognition achieves state-of-the-art 86% F-Score on ICDAR13
Real-time gesture recognition using eigenvectors
This paper discusses an implementation for gesture
recognition using eigenvectors under controlled conditions. This
application of eigenvector recognition is trained on a set of
defined hand images. Training images are processed using eigen
techniques from the OpenCV image processing library. Test images are then compared in real-time. These techniques are
outlined below.Telkom. CISCO, THRI
Scatteract: Automated extraction of data from scatter plots
Charts are an excellent way to convey patterns and trends in data, but they
do not facilitate further modeling of the data or close inspection of
individual data points. We present a fully automated system for extracting the
numerical values of data points from images of scatter plots. We use deep
learning techniques to identify the key components of the chart, and optical
character recognition together with robust regression to map from pixels to the
coordinate system of the chart. We focus on scatter plots with linear scales,
which already have several interesting challenges. Previous work has done fully
automatic extraction for other types of charts, but to our knowledge this is
the first approach that is fully automatic for scatter plots. Our method
performs well, achieving successful data extraction on 89% of the plots in our
test set.Comment: Submitted to ECML PKDD 2017 proceedings, 16 page
- …