143 research outputs found

    Implementation of Adaptive Unsharp Masking as a pre-filtering method for watermark detection and extraction

    Get PDF
    Digital watermarking has been one of the focal points of research interests in order to provide multimedia security in the last decade. Watermark data, belonging to the user, are embedded on an original work such as text, audio, image, and video and thus, product ownership can be proved. Various robust watermarking algorithms have been developed in order to extract/detect the watermark against such attacks. Although watermarking algorithms in the transform domain differ from others by different combinations of transform techniques, it is difficult to decide on an algorithm for a specific application. Therefore, instead of developing a new watermarking algorithm with different combinations of transform techniques, we propose a novel and effective watermark extraction and detection method by pre-filtering, namely Adaptive Unsharp Masking (AUM). In spite of the fact that Unsharp Masking (UM) based pre-filtering is used for watermark extraction/detection in the literature by causing the details of the watermarked image become more manifest, effectiveness of UM may decrease in some cases of attacks. In this study, AUM has been proposed for pre-filtering as a solution to the disadvantages of UM. Experimental results show that AUM performs better up to 11\% in objective quality metrics than that of the results when pre-filtering is not used. Moreover; AUM proposed for pre-filtering in the transform domain image watermarking is as effective as that of used in image enhancement and can be applied in an algorithm-independent way for pre-filtering in transform domain image watermarking

    Comparing Adobe’s Unsharp Masks and High-Pass Filters in Photoshop Using the Visual Information Fidelity Metric

    Get PDF
    The present study examines image sharpening techniques quantitatively. A technique known as unsharp masking has been the preferred image sharpening technique for imaging professionals for many years. More recently, another professional-level sharpening solution has been introduced, namely, the high-pass filter technique of image sharpening. An extensive review of the literature revealed no purely quantitative studies that compared these techniques. The present research compares unsharp masking (USM) and high-pass filter (HPF) sharpening using an image quality metric known as Visual Information Fidelity (VIF). Prior researchers have used VIF data in research aimed at improving the USM sharpening technique. The present study aims to add to this branch of the literature through the comparison of the USM and the HPF sharpening techniques. The objective of the present research is to determine which sharpening technique, USM or HPF, yields the highest VIF scores for two categories of images, macro images and architectural images. Each set of images was further analyzed to compare the VIF scores of subjects with high and low severity depth of field defects. Finally, the researcher proposed rules for choosing USM and HPF parameters that resulted in optimal VIF scores. For each category, the researcher captured 24 images (12 with high severity defects and 12 with low severity defects). Each image was sharpened using an iterative process of choosing USM and HPF sharpening parameters, applying sharpening filters with the chosen parameters, and assessing the resulting images using the VIF metric. The process was repeated until the VIF scores could no longer be improved. The highest USM and HPF VIF scores for each image were compared using a paired t-test for statistical significance. The t-test results demonstrated that: • The USM VIF scores for macro images (M = 1.86, SD = 0.59) outperformed those for HPF (M = 1.34, SD = 0.18), a statistically significant mean increase of 0.52, t = 5.57 (23), p = 0.0000115. Similar results were obtained for both the high severity and low severity subsets of macro images. • The USM VIF scores for architectural images (M = 1.40, SD = 0.24) outperformed those for HPF (M = 1.26, SD = 0.15), a statistically significant mean increase of 0.14, t = 5.21 (23), p = 0.0000276. Similar results were obtained for both the high severity and low severity subsets of architectural images. The researcher found that the optimal sharpening parameters for USM and HPF depend on the content of the image. The optimal choice of parameters for USM depends on whether the most important features are edges or objects. Specific rules for choosing USM parameters were developed for each class of images. HPF is simpler in the fact that it only uses one parameter, Radius. Specific rules for choosing the HPF Radius were also developed for each class of images. Based on these results, the researcher concluded that USM outperformed HPF in sharpening macro and architectural images. The superior performance of USM could be due to the fact that it provides more parameters for users to control the sharpening process than HPF

    Blockwise Transform Image Coding Enhancement and Edge Detection

    Get PDF
    The goal of this thesis is high quality image coding, enhancement and edge detection. A unified approach using novel fast transforms is developed to achieve all three objectives. Requirements are low bit rate, low complexity of implementation and parallel processing. The last requirement is achieved by processing the image in small blocks such that all blocks can be processed simultaneously. This is similar to biological vision. A major issue is to minimize the resulting block effects. This is done by using proper transforms and possibly an overlap-save technique. The bit rate in image coding is minimized by developing new results in optimal adaptive multistage transform coding. Newly developed fast trigonometric transforms are also utilized and compared for transform coding, image enhancement and edge detection. Both image enhancement and edge detection involve generalised bandpass filtering wit fast transforms. The algorithms have been developed with special attention to the properties of biological vision systems

    Development of Some Spatial-domain Preprocessing and Post-processing Algorithms for Better 2-D Up-scaling

    Get PDF
    Image super-resolution is an area of great interest in recent years and is extensively used in applications like video streaming, multimedia, internet technologies, consumer electronics, display and printing industries. Image super-resolution is a process of increasing the resolution of a given image without losing its integrity. Its most common application is to provide better visual effect after resizing a digital image for display or printing. One of the methods of improving the image resolution is through the employment of a 2-D interpolation. An up-scaled image should retain all the image details with very less degree of blurring meant for better visual quality. In literature, many efficient 2-D interpolation schemes are found that well preserve the image details in the up-scaled images; particularly at the regions with edges and fine details. Nevertheless, these existing interpolation schemes too give blurring effect in the up-scaled images due to the high frequency (HF) degradation during the up-sampling process. Hence, there is a scope to further improve their performance through the incorporation of various spatial domain pre-processing, post-processing and composite algorithms. Therefore, it is felt that there is sufficient scope to develop various efficient but simple pre-processing, post-processing and composite schemes to effectively restore the HF contents in the up-scaled images for various online and off-line applications. An efficient and widely used Lanczos-3 interpolation is taken for further performance improvement through the incorporation of various proposed algorithms. The various pre-processing algorithms developed in this thesis are summarized here. The term pre-processing refers to processing the low-resolution input image prior to image up-scaling. The various pre-processing algorithms proposed in this thesis are: Laplacian of Laplacian based global pre-processing (LLGP) scheme; Hybrid global pre-processing (HGP); Iterative Laplacian of Laplacian based global pre-processing (ILLGP); Unsharp masking based pre-processing (UMP); Iterative unsharp masking (IUM); Error based up-sampling(EU) scheme. The proposed algorithms: LLGP, HGP and ILLGP are three spatial domain preprocessing algorithms which are based on 4th, 6th and 8th order derivatives to alleviate nonuniform blurring in up-scaled images. These algorithms are used to obtain the high frequency (HF) extracts from an image by employing higher order derivatives and perform precise sharpening on a low resolution image to alleviate the blurring in its 2-D up-sampled counterpart. In case of unsharp masking based pre-processing (UMP) scheme, the blurred version of a low resolution image is used for HF extraction from the original version through image subtraction. The weighted version of the HF extracts are superimposed with the original image to produce a sharpened image prior to image up-scaling to counter blurring effectively. IUM makes use of many iterations to generate an unsharp mask which contains very high frequency (VHF) components. The VHF extract is the result of signal decomposition in terms of sub-bands using the concept of analysis filter bank. Since the degradation of VHF components is maximum, restoration of such components would produce much better restoration performance. EU is another pre-processing scheme in which the HF degradation due to image upscaling is extracted and is called prediction error. The prediction error contains the lost high frequency components. When this error is superimposed on the low resolution image prior to image up-sampling, blurring is considerably reduced in the up-scaled images. Various post-processing algorithms developed in this thesis are summarized in following. The term post-processing refers to processing the high resolution up-scaled image. The various post-processing algorithms proposed in this thesis are: Local adaptive Laplacian (LAL); Fuzzy weighted Laplacian (FWL); Legendre functional link artificial neural network(LFLANN). LAL is a non-fuzzy, local based scheme. The local regions of an up-scaled image with high variance are sharpened more than the region with moderate or low variance by employing a local adaptive Laplacian kernel. The weights of the LAL kernel are varied as per the normalized local variance so as to provide more degree of HF enhancement to high variance regions than the low variance counterpart to effectively counter the non-uniform blurring. Furthermore, FWL post-processing scheme with a higher degree of non-linearity is proposed to further improve the performance of LAL. FWL, being a fuzzy based mapping scheme, is highly nonlinear to resolve the blurring problem more effectively than LAL which employs a linear mapping. Another LFLANN based post-processing scheme is proposed here to minimize the cost function so as to reduce the blurring in a 2-D up-scaled image. Legendre polynomials are used for functional expansion of the input pattern-vector and provide high degree of nonlinearity. Therefore, the requirement of multiple layers can be replaced by single layer LFLANN architecture so as to reduce the cost function effectively for better restoration performance. With single layer architecture, it has reduced the computational complexity and hence is suitable for various real-time applications. There is a scope of further improvement of the stand-alone pre-processing and postprocessing schemes by combining them through composite schemes. Here, two spatial domain composite schemes, CS-I and CS-II are proposed to tackle non-uniform blurring in an up-scaled image. CS-I is developed by combining global iterative Laplacian (GIL) preprocessing scheme with LAL post-processing scheme. Another highly nonlinear composite scheme, CS-II is proposed which combines ILLGP scheme with a fuzzy weighted Laplacian post-processing scheme for more improved performance than the stand-alone schemes. Finally, it is observed that the proposed algorithms: ILLGP, IUM, FWL, LFLANN and CS-II are better algorithms in their respective categories for effectively reducing blurring in the up-scaled images

    Development of Impulsive Noise Detection Schemes for Selective Filtering in Images

    Get PDF
    Image Noise Suppression is a highly demanded approach in digital imaging systems design. Impulsive noise is one such noise, which is frequently encountered problem in acquistion, transmission and processing of images. In the area of image restoration, many state-of-the art filters consist of two main processes, classification (detection) and reconstruction (filtering). Classification is used to separate uncorrupted pixels from corrupted pixels. Reconstruction involves replacing the corrupted pixels by certain approximation technique. In this thesis such schemes of impulsive noise detection and filtering thereof are proposed. Impulsive noise can be Salt & Pepper Noise (SPN) or Random Valued Impulsive Noise (RVIN). Only RVIN model is considered in this thesis because of its realistic presence. In the RVIN model a corrupted pixel can take any value in the valid range. Adaptive threshold selection is emphasized for all the four proposed noise detection schemes. Incorporation of adaptive threshold into the noise detection process led to more reliable and more efficient detection of noise. Based on the noisy image characteristics and their statistics, threshold values are selected. To validate the efficacy of proposed noise filtering schemes, an application to image sharpening has been investigated under the noise conditions. It has been observed, if the noisy image passes through the sharpening scheme, the noise gets amplified and as a result the restored results are distorted. However, the prefiltering operations using the proposed schemes enhances the result to a greater extent. Extensive simulations and comparisons are done with competent schemes. It is observed, in general, that the proposed schemes are better in suppressing impulsive noise at different noise ratios than their counterparts

    Tissue segmentation using medical image processing chain optimization

    Get PDF
    Surveyed literature shows many segmentation algorithms using different types of optimization methods. These methods were used to minimize or maximize objective functions of entropy, similarity, clustering, contour, or thresholding. These specially developed functions target specific feature or step in the presented segmentation algorithms. To the best of our knowledge, this thesis is the first research work that uses an optimizer to build and optimize parameters of a full sequence of image processing chain. This thesis presents a universal algorithm that uses three images and their corresponding gold images to train the framework. The optimization algorithm explores the search space for the best sequence of the image processing chain to segment the targeted feature. Experiments indicate that using differential evolution to build Image processing chain (IPC) out of forty-five algorithms increases the segmentation performance of basic thresholding algorithms ranging from 2% to 78%

    Noise-Enhanced and Human Visual System-Driven Image Processing: Algorithms and Performance Limits

    Get PDF
    This dissertation investigates the problem of image processing based on stochastic resonance (SR) noise and human visual system (HVS) properties, where several novel frameworks and algorithms for object detection in images, image enhancement and image segmentation as well as the method to estimate the performance limit of image segmentation algorithms are developed. Object detection in images is a fundamental problem whose goal is to make a decision if the object of interest is present or absent in a given image. We develop a framework and algorithm to enhance the detection performance of suboptimal detectors using SR noise, where we add a suitable dose of noise into the original image data and obtain the performance improvement. Micro-calcification detection is employed in this dissertation as an illustrative example. The comparative experiments with a large number of images verify the efficiency of the presented approach. Image enhancement plays an important role and is widely used in various vision tasks. We develop two image enhancement approaches. One is based on SR noise, HVS-driven image quality evaluation metrics and the constrained multi-objective optimization (MOO) technique, which aims at refining the existing suboptimal image enhancement methods. Another is based on the selective enhancement framework, under which we develop several image enhancement algorithms. The two approaches are applied to many low quality images, and they outperform many existing enhancement algorithms. Image segmentation is critical to image analysis. We present two segmentation algorithms driven by HVS properties, where we incorporate the human visual perception factors into the segmentation procedure and encode the prior expectation on the segmentation results into the objective functions through Markov random fields (MRF). Our experimental results show that the presented algorithms achieve higher segmentation accuracy than many representative segmentation and clustering algorithms available in the literature. Performance limit, or performance bound, is very useful to evaluate different image segmentation algorithms and to analyze the segmentability of the given image content. We formulate image segmentation as a parameter estimation problem and derive a lower bound on the segmentation error, i.e., the mean square error (MSE) of the pixel labels considered in our work, using a modified Cramér-Rao bound (CRB). The derivation is based on the biased estimator assumption, whose reasonability is verified in this dissertation. Experimental results demonstrate the validity of the derived bound
    corecore