4 research outputs found
BiNet:Degraded-Manuscript Binarization in Diverse Document Textures and Layouts using Deep Encoder-Decoder Networks
Handwritten document-image binarization is a semantic segmentation process to
differentiate ink pixels from background pixels. It is one of the essential
steps towards character recognition, writer identification, and script-style
evolution analysis. The binarization task itself is challenging due to the vast
diversity of writing styles, inks, and paper materials. It is even more
difficult for historical manuscripts due to the aging and degradation of the
documents over time. One of such manuscripts is the Dead Sea Scrolls (DSS)
image collection, which poses extreme challenges for the existing binarization
techniques. This article proposes a new binarization technique for the DSS
images using the deep encoder-decoder networks. Although the artificial neural
network proposed here is primarily designed to binarize the DSS images, it can
be trained on different manuscript collections as well. Additionally, the use
of transfer learning makes the network already utilizable for a wide range of
handwritten documents, making it a unique multi-purpose tool for binarization.
Qualitative results and several quantitative comparisons using both historical
manuscripts and datasets from handwritten document image binarization
competition (H-DIBCO and DIBCO) exhibit the robustness and the effectiveness of
the system. The best performing network architecture proposed here is a variant
of the U-Net encoder-decoders.Comment: 26 pages, 15 figures, 11 table
Image Enhancement for Scanned Historical Documents in the Presence of Multiple Degradations
Historical documents are treasured sources of information but typically suffer from problems with quality and degradation. Scanned images of historical documents suffer from difficulties due to paper quality and poor image capture, producing images with low contrast, smeared ink, bleed-through and uneven illumination. This PhD thesis proposes a novel adaptative histogram matching method to remove these artefacts from scanned images of historical documents. The adaptive histogram matching is modelled to create an ideal histogram by dividing the histogram using its Otsu level and applying Gaussian distributions to each segment with iterative output refinement applied to individual images. The pre-processing techniques of contrast stretching, wiener filtering, and bilateral filtering are used before the proposed adaptive histogram matching approach to maximise the dynamic range and reduce noise. The goal is to better represent document images and improve readability and the source images for Optical Character Recognition (OCR). Unlike other enhancement methods designed for single artefacts, the proposed method enhances multiple (low-contrast, smeared-ink, bleed-through and uneven illumination). In addition to developing an algorithm for historical document enhancement, the research also contributes a new dataset of scanned historical newspapers (an annotated subset of the Europeana Newspaper - ENP – dataset) where the enhancement technique is tested, which can also be used for further research. Experimental results show that the proposed method significantly reduces background noise and improves image quality on multiple artefacts compared to other enhancement methods. Several performance criteria are utilised to evaluate the proposed method’s efficiency. These include Signal to Noise Ratio (SNR), Mean opinion score (MOS), and visual document image quality assessment (VDIQA) metric called Visual Document Image Quality Assessment Metric (VDQAM). Additional assessment criteria to measure post-processing binarization quality are also discussed with enhanced results based on the Peak signal-to-noise ratio (PSNR), negative rate metric (NRM) and F-measure.Keywords: Image Enhancement, Historical Documents, OCR, Digitisation, Adaptive histogram matchin