377 research outputs found

    Image watermarking, steganography, and morphological processing

    Get PDF
    With the fast development of computer technology, research in the fields of multimedia security, image processing, and robot vision have recently become popular. Image watermarking, steganogrphic system, morphological processing and shortest path planning are important subjects among them. In this dissertation, the fundamental techniques are reviewed first followed by the presentation of novel algorithms and theorems for these three subjects. The research on multimedia security consists of two parts, image watermarking and steganographic system. In image watermarking, several algorithms are developed to achieve different goals as shown below. In order to embed more watermarks and to minimize distortion of watermarked images, a novel watermarking technique using combinational spatial and frequency domains is presented. In order to correct rounding errors, a novel technique based on the genetic algorithm (GA) is developed. By separating medical images into Region of Interest (ROI) and non-ROI parts, higher compression rates can be achieved where the ROI is compressed by lossless compression and the non-ROI by lossy compression. The GA-based watermarking technique can also be considered as a fundamental platform for other fragile watermarking techniques. In order to simplify the selection and integrate different watermarking techniques, a novel adjusted-purpose digital watermarking is developed. In order to enlarge the capacity of robust watermarking, a novel robust high-capacity watermarking is developed. In steganographic system, a novel steganographic algorithm is developed by using GA to break the inspection of steganalytic system. In morphological processing, the GA-based techniques are developed to decompose arbitrary shapes of big binary structuring elements and arbitrary values of big grayscale structuring elements into small ones. The decomposition is suited for a parallel-pipelined architecture. The techniques can speed up the morphological processing and allow full freedom for users to design any type and any size of binary and grayscale structuring elements. In applications such as shortest path planning, a novel method is first presented to obtaining Euclidean distance transformation (EDT) in just two scans of image. The shortest path can be extracted based on distance maps by tracking minimum values. In order to record the motion path, a new chain-code representation is developed to allow forward and backward movements. By placing the smooth turning-angle constraint, it is possible to mimic realistic motions of cars. By using dynamically rotational morphology, it is not only guarantee collision-free in the shortest path, but also reduce time complexity dramatically. As soon as the distance map of a destination and collision-free codes have been established off-line, shortest paths of cars given any starting location toward the destination can be promptly obtained on-line

    Study and Comparison of Surface Roughness Measurements

    No full text
    National audienceThis survey paper focus on recent researches whose goal is to optimize treatments on 3D meshes, thanks to a study of their surface features, and more precisely their roughness and saliency. Applications like watermarking or lossy compression can benefit from a precise roughness detection, to better hide the watermarks or quantize coarsely these areas, without altering visually the shape. Despite investigations on scale dependence leading to multi-scale approaches, an accurate roughness or pattern characterization is still lacking, but challenging for those treatments. We think there is still room for investigations that could benefit from the power of the wavelet analysis or the fractal models. Furthermore only few works are now able to differentiate roughness from saliency, though it is essential for faithfully simplifying or denoising a 3D mesh. Hence we have investigated roughness quantification methods for analog surfaces, in several domains of physics. Some roughness parameters used in these fields and the additionnal information they bring are finally studied, since we think an adaptation for 3D meshes could be beneficial

    Texture based Image Splicing Forgery Recognition using a Passive Approach

    Get PDF
    With the growing usage of the internet in daily life along with the usage of dominant picture editing software tools in creating forged pictures effortlessly, make us lose trust in the authenticity of the images. For more than a decade, extensive research is going on in the Image forensic area that aims at restoring trustworthiness in images by bringing various tampering detection techniques. In the proposed method, identification of image splicing technique is introduced which depends on the picture texture analysis which characterizes the picture areas by the content of the texture. In this method, an image is characterized by the regions of their texture content. The experimental outcomes describe that the proposed method is effective to identify spliced picture forgery with an accuracy of 79.5%

    Lossless and low-cost integer-based lifting wavelet transform

    Get PDF
    Discrete wavelet transform (DWT) is a powerful tool for analyzing real-time signals, including aperiodic, irregular, noisy, and transient data, because of its capability to explore signals in both the frequency- and time-domain in different resolutions. For this reason, they are used extensively in a wide number of applications in image and signal processing. Despite the wide usage, the implementation of the wavelet transform is usually lossy or computationally complex, and it requires expensive hardware. However, in many applications, such as medical diagnosis, reversible data-hiding, and critical satellite data, lossless implementation of the wavelet transform is desirable. It is also important to have more hardware-friendly implementations due to its recent inclusion in signal processing modules in system-on-chips (SoCs). To address the need, this research work provides a generalized implementation of a wavelet transform using an integer-based lifting method to produce lossless and low-cost architecture while maintaining the performance close to the original wavelets. In order to achieve a general implementation method for all orthogonal and biorthogonal wavelets, the Daubechies wavelet family has been utilized at first since it is one of the most widely used wavelets and based on a systematic method of construction of compact support orthogonal wavelets. Though the first two phases of this work are for Daubechies wavelets, they can be generalized in order to apply to other wavelets as well. Subsequently, some techniques used in the primary works have been adopted and the critical issues for achieving general lossless implementation have solved to propose a general lossless method. The research work presented here can be divided into several phases. In the first phase, low-cost architectures of the Daubechies-4 (D4) and Daubechies-6 (D6) wavelets have been derived by applying the integer-polynomial mapping. A lifting architecture has been used which reduces the cost by a half compared to the conventional convolution-based approach. The application of integer-polynomial mapping (IPM) of the polynomial filter coefficient with a floating-point value further decreases the complexity and reduces the loss in signal reconstruction. Also, the “resource sharing” between lifting steps results in a further reduction in implementation costs and near-lossless data reconstruction. In the second phase, a completely lossless or error-free architecture has been proposed for the Daubechies-8 (D8) wavelet. Several lifting variants have been derived for the same wavelet, the integer mapping has been applied, and the best variant is determined in terms of performance, using entropy and transform coding gain. Then a theory has been derived regarding the impact of scaling steps on the transform coding gain (GT). The approach results in the lowest cost lossless architecture of the D8 in the literature, to the best of our knowledge. The proposed approach may be applied to other orthogonal wavelets, including biorthogonal ones to achieve higher performance. In the final phase, a general algorithm has been proposed to implement the original filter coefficients expressed by a polyphase matrix into a more efficient lifting structure. This is done by using modified factorization, so that the factorized polyphase matrix does not include the lossy scaling step like the conventional lifting method. This general technique has been applied on some widely used orthogonal and biorthogonal wavelets and its advantages have been discussed. Since the discrete wavelet transform is used in a vast number of applications, the proposed algorithms can be utilized in those cases to achieve lossless, low-cost, and hardware-friendly architectures

    Signal processing algorithms for enhanced image fusion performance and assessment

    Get PDF
    The dissertation presents several signal processing algorithms for image fusion in noisy multimodal conditions. It introduces a novel image fusion method which performs well for image sets heavily corrupted by noise. As opposed to current image fusion schemes, the method has no requirements for a priori knowledge of the noise component. The image is decomposed with Chebyshev polynomials (CP) being used as basis functions to perform fusion at feature level. The properties of CP, namely fast convergence and smooth approximation, renders it ideal for heuristic and indiscriminate denoising fusion tasks. Quantitative evaluation using objective fusion assessment methods show favourable performance of the proposed scheme compared to previous efforts on image fusion, notably in heavily corrupted images. The approach is further improved by incorporating the advantages of CP with a state-of-the-art fusion technique named independent component analysis (ICA), for joint-fusion processing based on region saliency. Whilst CP fusion is robust under severe noise conditions, it is prone to eliminating high frequency information of the images involved, thereby limiting image sharpness. Fusion using ICA, on the other hand, performs well in transferring edges and other salient features of the input images into the composite output. The combination of both methods, coupled with several mathematical morphological operations in an algorithm fusion framework, is considered a viable solution. Again, according to the quantitative metrics the results of our proposed approach are very encouraging as far as joint fusion and denoising are concerned. Another focus of this dissertation is on a novel metric for image fusion evaluation that is based on texture. The conservation of background textural details is considered important in many fusion applications as they help define the image depth and structure, which may prove crucial in many surveillance and remote sensing applications. Our work aims to evaluate the performance of image fusion algorithms based on their ability to retain textural details from the fusion process. This is done by utilising the gray-level co-occurrence matrix (GLCM) model to extract second-order statistical features for the derivation of an image textural measure, which is then used to replace the edge-based calculations in an objective-based fusion metric. Performance evaluation on established fusion methods verifies that the proposed metric is viable, especially for multimodal scenarios

    A review on copy-move image forgery detection techniques

    Get PDF
    With billions of digital images flooding the internet which are widely used and regards as the major information source in many fields in recent years. With the high advance of technology, it may seem easy to fraud the image. In digital images, copy-move forgery is the most common image tampering, where some object(s) or region(s) duplicate in the digital image. The important research has attracted more attention in digital forensic is forgery detection and localization. Many techniques have been proposed and many papers have been published to detect image forgery. This paper introduced a review of research papers on copy-move image forgery published in reputed journals from 2017 to 2020 and focused on discussing various strategies related with fraud images to highlight on the latest tools used in the detection. This article will help the researchers to understand the current algorithms and techniques in this field and ultimately develop new and more efficient algorithms of detection copy-move image
    • …
    corecore