4 research outputs found

    Text area detection in handwritten documents scanned for further processing

    No full text
    In this paper we present an approach to text area detection using binary images, Constrained Run Length Algorithm and other noise reduction methods of removing the artefacts. Text processing includes various activities, most of which are related to preparing input data for further operations in the best possible way, that will not hinder the OCR algorithms. This is especially the case when handwritten manuscripts are considered, and even more so with very old documents. We present our methodology for text area detection problem, which is capable of removing most of irrelevant objects, including elements such as page edges, stains, folds etc. At the same time the presented method can handle multi-column texts or varying line thickness. The generated mask can accurately mark the actual text area, so that the output image can be easily used in further text processing steps

    Textural features based on run length encoding in the classification of furniture surfaces with the orange skin defect

    No full text
    Textural features based upon thresholding and run length encoding have been successfully applied to the problem of classification of the quality of lacquered surfaces in furniture exhibiting the surface defect known as orange skin. The set of features for one surface patch consists of 12 real numbers. The classifier used was the one nearest neighbour classifier without feature selection. The classification quality was tested on 808 images 300 by 300 pixels, made under controlled, close-to-tangential lighting, with three classes: good, acceptable and bad, in close to balanced numbers. The classification accuracy was not smaller than 98% when the tested surface was not rotated with respect to the training samples, 97% for rotations up to 20 degrees and 95.5% in the worst case for arbitrary rotations

    Context-based segmentation of the longissimus muscle in beef with a deep neural network

    No full text
    The problem of segmenting the cross-section through the longissimus muscle in beef carcasses with computer vision methods was investigated. The available data were 111 images of cross-sections coming from 28 cows (typically four images per cow). Training data were the pixels of the muscles, marked manually. The AlexNet deep convolutional neural network was used as the classifier, and single pixels were the classified objects. Each pixel was presented to the network together with its small circular neighbourhood, and with its context represented by the further neighbourhood, darkened by halving the image intensity. The average classification accuracy was 96%. The accuracy without darkening the context was found to be smaller, with a small but statistically significant difference. The segmentation of the longissimus muscle is the introductory stage for the next steps of assessing the quality of beef for the alimentary purposes
    corecore