772 research outputs found

    Robust Adaptive Median Binary Pattern for noisy texture classification and retrieval

    Full text link
    Texture is an important cue for different computer vision tasks and applications. Local Binary Pattern (LBP) is considered one of the best yet efficient texture descriptors. However, LBP has some notable limitations, mostly the sensitivity to noise. In this paper, we address these criteria by introducing a novel texture descriptor, Robust Adaptive Median Binary Pattern (RAMBP). RAMBP based on classification process of noisy pixels, adaptive analysis window, scale analysis and image regions median comparison. The proposed method handles images with high noisy textures, and increases the discriminative properties by capturing microstructure and macrostructure texture information. The proposed method has been evaluated on popular texture datasets for classification and retrieval tasks, and under different high noise conditions. Without any train or prior knowledge of noise type, RAMBP achieved the best classification compared to state-of-the-art techniques. It scored more than 90%90\% under 50%50\% impulse noise densities, more than 95%95\% under Gaussian noised textures with standard deviation σ=5\sigma = 5, and more than 99%99\% under Gaussian blurred textures with standard deviation σ=1.25\sigma = 1.25. The proposed method yielded competitive results and high performance as one of the best descriptors in noise-free texture classification. Furthermore, RAMBP showed also high performance for the problem of noisy texture retrieval providing high scores of recall and precision measures for textures with high levels of noise

    Study Of Gaussian & Impulsive Noise Suppression Schemes In Images

    Get PDF
    Noise is introduced into images usually while transferring and acquiring them.The main type of noise added while image acquisition is called Gaussian noise while Impulsive noise is generally introduced while transmitting image data over an unsecure communication channel , while it can also be added by acquiring. Gaussian noise is a set of values taken from a zero mean Gaussian distribution which are added to each pixel value. Impulsive noise involves changing a part of the pixel values with random ones. Various techniques are employed for the removal of these types of noise based on the properties of their respective noise models. Impulse Noise removal algorithms popularly use ordered statistics based ¯lters. The ¯rst one is an adaptive ¯lter using center-weighted median. In this method, the di®erence of the center weighted mean of a neighborhood with the central pixel under consideration is compared with a set of thresholds. Another method which takes into account the presence of the noise free pixels has been implemented.It convolutes the median of each neighborhood with a set of convolution kernels which are oriented according to all possible con¯gurations of edges that contain the central pixel,if it lies on an edge. A third method which deals with the detection of noisy pixels on the binary slices of an image is implemented. It is based on threshold Boolean ¯ltering. The ¯lter inverts the value of the central pixel if the number of pixels with values opposite to it is more than the threshold. The fourth method has an e±cient double derivative detector, which gives a de- cision based on the value of the double derivative. The substitution is done with the average gray scale value of the neighborhood. Gaussian Noise removal algorithms ideally should smooth the distinct parts of the image without blurring the edges.A universal noise removing scheme is implemented which weighs each pixel with respect to its neighborhood and deals with Gaussian and impulse noise pixels di®erently based on parameter values for spatial, radiometric and impulsive weight of the central pixel. The aforementioned techniques are implemented and their results are compared subjectively as well as objectively

    Adaptive Algorithms for Automated Processing of Document Images

    Get PDF
    Large scale document digitization projects continue to motivate interesting document understanding technologies such as script and language identification, page classification, segmentation and enhancement. Typically, however, solutions are still limited to narrow domains or regular formats such as books, forms, articles or letters and operate best on clean documents scanned in a controlled environment. More general collections of heterogeneous documents challenge the basic assumptions of state-of-the-art technology regarding quality, script, content and layout. Our work explores the use of adaptive algorithms for the automated analysis of noisy and complex document collections. We first propose, implement and evaluate an adaptive clutter detection and removal technique for complex binary documents. Our distance transform based technique aims to remove irregular and independent unwanted foreground content while leaving text content untouched. The novelty of this approach is in its determination of best approximation to clutter-content boundary with text like structures. Second, we describe a page segmentation technique called Voronoi++ for complex layouts which builds upon the state-of-the-art method proposed by Kise [Kise1999]. Our approach does not assume structured text zones and is designed to handle multi-lingual text in both handwritten and printed form. Voronoi++ is a dynamically adaptive and contextually aware approach that considers components' separation features combined with Docstrum [O'Gorman1993] based angular and neighborhood features to form provisional zone hypotheses. These provisional zones are then verified based on the context built from local separation and high-level content features. Finally, our research proposes a generic model to segment and to recognize characters for any complex syllabic or non-syllabic script, using font-models. This concept is based on the fact that font files contain all the information necessary to render text and thus a model for how to decompose them. Instead of script-specific routines, this work is a step towards a generic character and recognition scheme for both Latin and non-Latin scripts

    A Novel DWT-CT approach in Digital Watermarking using PSO

    Get PDF
    The importance of watermarking is dramatically enhanced due to the promising technologies like Internet of Things (IoT), Data analysis, and automation of identification in many sectors. Due to these reasons, systems are inter-connected through networking and internet and huge amounts of information is generated, distributed and transmitted over the World Wide Web. Thus authentication of the information is a challenging task. The algorithm developed for the watermarking needs to be robust against various attack such as salt & peppers, filtering, compression and cropping etc. This paper focuses on the robustness of the algorithm by using a hybrid approach of two transforms such as Contourlet, Discrete Wavelet Transform (DWT). Also, the Particle Swarm Optimization (PSO) is used to optimize the embedding strength factor. The proposed digital watermarking algorithm has been tested against common types of image attacks. Experiment results for the proposed algorithm gives better performance by using similarity metrics such as NCC (Normalized Cross Correlation value) and PSNR (Peak Signal to Noise Ratio)

    Print-Scan Resilient Text Image Watermarking Based on Stroke Direction Modulation for Chinese Document Authentication

    Get PDF
    Print-scan resilient watermarking has emerged as an attractive way for document security. This paper proposes an stroke direction modulation technique for watermarking in Chinese text images. The watermark produced by the idea offers robustness to print-photocopy-scan, yet provides relatively high embedding capacity without losing the transparency. During the embedding phase, the angle of rotatable strokes are quantized to embed the bits. This requires several stages of preprocessing, including stroke generation, junction searching, rotatable stroke decision and character partition. Moreover, shuffling is applied to equalize the uneven embedding capacity. For the data detection, denoising and deskewing mechanisms are used to compensate for the distortions induced by hardcopy. Experimental results show that our technique attains high detection accuracy against distortions resulting from print-scan operations, good quality photocopies and benign attacks in accord with the future goal of soft authentication

    Entropy Based Robust Watermarking Algorithm

    Get PDF
    Tänu aina kasvavale multimeedia andmeedastus mahtudele Internetis, on esile kerkinud mured turvalisusest ja piraatlusest. Digitaalse meedia paljundamise ja muutmise maht on loonud vajaduse digitaalse meedia vesimärgistamise järgi. Selles töös on tutvustatud vastupidavaid vesimärkide lisamise algoritme, mis lisavad vesimärgid madala entroopiaga pildi osadesse. Välja pakutud algoritmides jagatakse algne pilt blokkidesse ning arvutatakse iga bloki entroopia. Kõikide blokkide keskmine entroopia väärtus valitakse künniseks, mille järgi otsustatakse, millistesse blokkidesse vesimärk lisada. Kõik blokid, mille entroopia on väiksem kui künnis, viiakse signaali sageduse kujule kasutades Discrete Wavelet Transform algoritmi. Madala sagedusega sagedusvahemikule rakendatakse Chirp Z-Transform algoritmi ja saadud tulemusele LU-dekompositsiooni või QR-dekompositsiooni. Singular Value Decomposition meetodi rakendamisel diagonaalmaatriksile, mis saadi eelmisest sammust, saadakse iga bloki vastav väärtus. Vesimärk lisatakse pildile, liites iga bloki arvutatud väärtusele vesimärgi Singular Value Decomposition meetodi tulemused. Kirjeldatud algoritme testiti ning võrreldi teiste tavapärast ning uudsete vesimärkide lisamise tehnoloogiatega. Kvantitatiivsed ja kvalitatiivsed eksperimendid näitavad, et välja pakutud meetodid on tajumatud ning vastupidavad signaali töötlemise rünnakutele.With growth of digital media distributed over the Internet, concerns about security and piracy have emerged. The amount of digital media reproduction and tampering has brought a need for content watermarking. In this work, multiple robust watermarking algorithms are introduced. They embed watermark image into singular values of host image’s blocks with low entropy values. In proposed algorithms, host image is divided into blocks, and the entropy of each block is calculated. The average of all entropies indicates the chosen threshold value for selecting the blocks in which watermark image should be embedded. All blocks with entropy lower than the calculated threshold are decomposed into frequency subbands using discrete wavelet transform (DWT). Subsequently chirp z-transform (CZT) is applied to the low-frequency subband followed by an appropriate matrix decomposition such as lower and upper decomposition (LUD) or orthogonal-triangular decomposition (QR decomposition). By applying singular value decomposition (SVD) to diagonal matrices obtained by the aforementioned matrix decompositions, the singular values of each block are calculated. Watermark image is embedded by adding singular values of the watermark image to singular values of the low entropy blocks. Proposed algorithms are tested on many host and watermark images, and they are compared with conventional and other state-of-the-art watermarking techniques. The quantitative and qualitative experimental results are indicating that the proposed algorithms are imperceptible and robust against many signal processing attacks

    Digits Recognition on Medical Device

    Get PDF
    With the rapid development of mobile health, mechanisms for automatic data input are becoming increasingly important for mobile health apps. In these apps, users are often required to input data frequently, especially numbers, from medical devices such as glucometers and blood pressure meters. However, these simple tasks are tedious and prone to error. Even though some Bluetooth devices can make those input operations easier, they are not popular enough due to being expensive and requiring complicated protocol support. Therefore, we propose an automatic procedure to recognize the digits on the screen of medical devices with smartphone cameras. The whole procedure includes several “standard” components in computer vision: image enhancement, the region-of-interest detection, and text recognition. Previous works existed for each component, but they have various weaknesses that lead to a low recognition rate. We proposed several novel enhancements in each component. Experiment results suggest that our enhanced procedure outperforms the procedure of applying optical character recognition directly from 6.2% to 62.1%. This procedure can be adopted (with human verification) to recognize the digits on the screen of medical devices with smartphone cameras

    Quantitative Evaluation of Dense Skeletons for Image Compression

    Get PDF
    Skeletons are well-known descriptors used for analysis and processing of 2D binary images. Recently, dense skeletons have been proposed as an extension of classical skeletons as a dual encoding for 2D grayscale and color images. Yet, their encoding power, measured by the quality and size of the encoded image, and how these metrics depend on selected encoding parameters, has not been formally evaluated. In this paper, we fill this gap with two main contributions. First, we improve the encoding power of dense skeletons by effective layer selection heuristics, a refined skeleton pixel-chain encoding, and a postprocessing compression scheme. Secondly, we propose a benchmark to assess the encoding power of dense skeletons for a wide set of natural and synthetic color and grayscale images. We use this benchmark to derive optimal parameters for dense skeletons. Our method, called Compressing Dense Medial Descriptors (CDMD), achieves higher-compression ratios at similar quality to the well-known JPEG technique and, thereby, shows that skeletons can be an interesting option for lossy image encoding

    A robust region-adaptive digital image watermarking system

    Get PDF
    Digital image watermarking techniques have drawn the attention of researchers and practitioners as a means of protecting copyright in digital images. The technique involves a subset of information-hiding technologies, which work by embedding information into a host image without perceptually altering the appearance of the host image. Despite progress in digital image watermarking technology, the main objectives of the majority of research in this area remain improvements in the imperceptibility and robustness of the watermark to attacks. Watermark attacks are often deliberately applied to a watermarked image in order to remove or destroy any watermark signals in the host data. The purpose of the attack is. aimed at disabling the copyright protection system offered by watermarking technology. Our research in the area of watermark attacks found a number of different types, which can be classified into a number of categories including removal attacks, geometry attacks, cryptographic attacks and protocol attacks. Our research also found that both pixel domain and transform domain watermarking techniques share similar levels of sensitivity to these attacks. The experiment conducted to analyse the effects of different attacks on watermarked data provided us with the conclusion that each attack affects the high and low frequency part of the watermarked image spectrum differently. Furthermore, the findings also showed that the effects of an attack can be alleviated by using a watermark image with a similar frequency spectrum to that of the host image. The results of this experiment led us to a hypothesis that would be proven by applying a watermark embedding technique which takes into account all of the above phenomena. We call this technique 'region-adaptive watermarking'. Region-adaptive watermarking is a novel embedding technique where the watermark data is embedded in different regions of the host image. The embedding algorithms use discrete wavelet transforms and a combination of discrete wavelet transforms and singular value decomposition, respectively. This technique is derived from the earlier hypothesis that the robustness of a watermarking process can be improved by using watermark data in the frequency spectrum that are not too dissimilar to that of the host data. To facilitate this, the technique utilises dual watermarking technologies and embeds parts of the watermark images into selected regions of the host image. Our experiment shows that our technique improves the robustness of the watermark data to image processing and geometric attacks, thus validating the earlier hypothesis. In addition to improving the robustness of the watermark to attacks, we can also show a novel use for the region-adaptive watermarking technique as a means of detecting whether certain types of attack have occurred. This is a unique feature of our watermarking algorithm, which separates it from other state-of-the-art techniques. The watermark detection process uses coefficients derived from the region-adaptive watermarking algorithm in a linear classifier. The experiment conducted to validate this feature shows that, on average, 94.5% of all watermark attacks can be correctly detected and identified
    corecore