687 research outputs found
Persian Heritage Image Binarization Competition (PHIBC 2012)
The first competition on the binarization of historical Persian documents and
manuscripts (PHIBC 2012) has been organized in conjunction with the first
Iranian conference on pattern recognition and image analysis (PRIA 2013). The
main objective of PHIBC 2012 is to evaluate performance of the binarization
methodologies, when applied on the Persian heritage images. This paper provides
a report on the methodology and performance of the three submitted algorithms
based on evaluation measures has been used.Comment: 4 pages, 2 figures, conferenc
Learning Surrogate Models of Document Image Quality Metrics for Automated Document Image Processing
Computation of document image quality metrics often depends upon the
availability of a ground truth image corresponding to the document. This limits
the applicability of quality metrics in applications such as hyperparameter
optimization of image processing algorithms that operate on-the-fly on unseen
documents. This work proposes the use of surrogate models to learn the behavior
of a given document quality metric on existing datasets where ground truth
images are available. The trained surrogate model can later be used to predict
the metric value on previously unseen document images without requiring access
to ground truth images. The surrogate model is empirically evaluated on the
Document Image Binarization Competition (DIBCO) and the Handwritten Document
Image Binarization Competition (H-DIBCO) datasets
Recognizing Degraded Handwritten Characters
In this paper, Slavonic manuscripts from the 11th
century written in Glagolitic script are
investigated. State-of-the-art optical character recognition methods produce poor results
for degraded handwritten document images. This is largely due to a lack of suitable
results from basic pre-processing steps such as binarization and image segmentation.
Therefore, a new, binarization-free approach will be presented that is independent of
pre-processing deficiencies. It additionally incorporates local information in order to
recognize also fragmented or faded characters. The proposed algorithm consists of
two steps: character classification and character localization. Firstly scale invariant
feature transform features are extracted and classified using support vector machines.
On this basis interest points are clustered according to their spatial information. Then,
characters are localized and eventually recognized by a weighted voting scheme of
pre-classified local descriptors. Preliminary results show that the proposed system can
handle highly degraded manuscript images with background noise, e.g. stains, tears,
and faded characters
A Multiple-Expert Binarization Framework for Multispectral Images
In this work, a multiple-expert binarization framework for multispectral
images is proposed. The framework is based on a constrained subspace selection
limited to the spectral bands combined with state-of-the-art gray-level
binarization methods. The framework uses a binarization wrapper to enhance the
performance of the gray-level binarization. Nonlinear preprocessing of the
individual spectral bands is used to enhance the textual information. An
evolutionary optimizer is considered to obtain the optimal and some suboptimal
3-band subspaces from which an ensemble of experts is then formed. The
framework is applied to a ground truth multispectral dataset with promising
results. In addition, a generalization to the cross-validation approach is
developed that not only evaluates generalizability of the framework, it also
provides a practical instance of the selected experts that could be then
applied to unseen inputs despite the small size of the given ground truth
dataset.Comment: 12 pages, 8 figures, 6 tables. Presented at ICDAR'1
Image Enhancement with Statistical Estimation
Contrast enhancement is an important area of research for the image analysis.
Over the decade, the researcher worked on this domain to develop an efficient
and adequate algorithm. The proposed method will enhance the contrast of image
using Binarization method with the help of Maximum Likelihood Estimation (MLE).
The paper aims to enhance the image contrast of bimodal and multi-modal images.
The proposed methodology use to collect mathematical information retrieves from
the image. In this paper, we are using binarization method that generates the
desired histogram by separating image nodes. It generates the enhanced image
using histogram specification with binarization method. The proposed method has
showed an improvement in the image contrast enhancement compare with the other
image.Comment: 9 pages,6 figures; ISSN:0975-5578 (Online); 0975-5934 (Print
Application of Threshold Techniques for Readability Improvement of Jawi Historical Manuscript Images
Historical documents such as old books and manuscripts have a high aesthetic
value and highly appreciated. Unfortunately, there are some documents cannot be
read due to quality problems like faded paper, ink expand, uneven colour tone,
torn paper and other elements disruption such as the existence of small spots.
The study aims to produce a copy of manuscript that shows clear wordings so
they can easily be read and the copy can also be displayed for visitors. 16
samples of Jawi historical manuscript with different quality problems were
obtained from The Royal Museum of Pahang, Malaysia. We applied three
binarization techniques; Otsu's method represents global threshold technique;
Sauvola and Niblack method which are categorized as local threshold techniques.
We compared the binarized images with the original manuscript to be visually
inspected by the museum's curator. The unclear features were marked and
analyzed. Most of the examined images show that with optimal parameters and
effective pre processing technique, local thresholding methods are work well
compare with the other one. Niblack's and Sauvola's techniques seem to be the
suitable approaches for these types of images. Most of binarized images with
these two methods show improvement for readability and character recognition.
For this research, even the differences of image result were hard to be
distinguished by human capabilities, after comparing the time cost and overall
achievement rate of recognized symbols, Niblack's method is performing better
than Sauvola's. We could improve the post processing step by adding edge
detection techniques and further enhanced by an innovative image refinement
technique and a formulation of a class proper method.Comment: 10 pages, 6 figures, 2 tables, Advance Computing: An International
Journal (ACIJ
- …