17 research outputs found

    An Automated Algorithm for Approximation of Temporal Video Data Using Linear B'EZIER Fitting

    Full text link
    This paper presents an efficient method for approximation of temporal video data using linear Bezier fitting. For a given sequence of frames, the proposed method estimates the intensity variations of each pixel in temporal dimension using linear Bezier fitting in Euclidean space. Fitting of each segment ensures upper bound of specified mean squared error. Break and fit criteria is employed to minimize the number of segments required to fit the data. The proposed method is well suitable for lossy compression of temporal video data and automates the fitting process of each pixel. Experimental results show that the proposed method yields good results both in terms of objective and subjective quality measurement parameters without causing any blocking artifacts.Comment: 14 Pages, IJMA 201

    Blind colour image watermarking techniques in hybrid domain using least significant bit and slantlet transform

    Get PDF
    Colour image watermarking has attracted a lot of interests since the last decade in tandem with the rapid growth of internet and its applications. This is due to increased awareness especially amongst netizens to protect digital assets from fraudulent activities. Many research efforts focused on improving the imperceptibility or robustness of both semi-blind and non-blind watermarking in spatial or transform domain. The results so far have been encouraging. Nonetheless, the requirements of the watermarking applications are varied in terms of imperceptibility, robustness and capacity. Ironically, limited studies concern on the authenticity and blind watermarking. Hence, this study presents two new blind RGB image watermarking techniques called Model1 and Model2 in hybrid domain using Least Significant Bit (LSB) insertion and Slantlet Transform (SLT). The models share similar pre-processing and LSB insertion stages but differ in SLT approach. In addition, two interrelated watermarks known as main watermark (MW) and sub-watermark (SW) are also utilized. Firstly, the RGB cover image is converted into YCbCr colour space and then split up into three components namely, Y, Cb and Cr. Secondly, the Cb component is selected as a cover for the MW embedding using the LSB substitution to attain a Cb-watermarked image (CbW). Thirdly, the Cr component is chosen and converted into the transform domain using SLT, and is subsequently decomposed into two paths: three-level sub-bands for Model1 and two-level sub-bands for Model2. For each model, the sub-bands are then used as a cover for sub-watermark embedding to generate a Cr-watermarked image (CrW). Following that, the Y component, CbW and CrW are combined to obtain a YCbCr-watermarked image. Finally, the image is reverted to RGB colour space to attain the actual watermarked image (WI). Upon embedding, the MW and SW are extracted from WI. The extraction process is similar to the above embedding except it is accomplished in a reverse order. Experimental results which utilized the standard dataset with fifteen well-known attacks revealed that, among others: Model1 has produced high imperceptibility, moderate robustness and good capacity, with Peak Signal-to-Noise Ratio (PSNR) rose to 65dB, Normalized Cross Correlation (NCC) moderated at 0.80, and capacity was 15%. Meanwhile, Model2, as per designed, performed positively in all aspects, with NCC strengthened to 1.00, capacity jumped to 25% and PSNR softened at 55dB but still on the high side. Interestingly, in terms of authenticity, Model2 performed impressively albeit the extracted MW has been completely altered. Overall, the models have successfully fulfilled all the research objectives and also markedly outperformed benchmark watermarking techniques

    DCT Implementation on GPU

    Get PDF
    There has been a great progress in the field of graphics processors. Since, there is no rise in the speed of the normal CPU processors; Designers are coming up with multi-core, parallel processors. Because of their popularity in parallel processing, GPUs are becoming more and more attractive for many applications. With the increasing demand in utilizing GPUs, there is a great need to develop operating systems that handle the GPU to full capacity. GPUs offer a very efficient environment for many image processing applications. This thesis explores the processing power of GPUs for digital image compression using Discrete cosine transform

    Analysis of Image Compression Methods Based On Transform and Fractal Coding

    Get PDF
    Image compression is process to remove the redundant information from the image so that only essential information can be stored to reduce the storage size, transmission bandwidth and transmission time. The essential information is extracted by various transforms techniques such that it can be reconstructed without losing quality and information of the image. In this thesis work comparative analysis of image compression is done by four transform method, which are Discrete Cosine Transform (DCT), Discrete Wavelet Transform( DWT) & Hybrid (DCT+DWT) Transform and fractal coding. MATLAB programs were written for each of the above method and concluded based on the results obtained that hybrid DWT-DCT algorithm performs much better than the standalone JPEG-based DCT, DWT algorithms in terms of peak signal to noise ratio (PSNR), as well as visual perception at higher compression ratio. The popular JPEG standard is widely used in digital cameras and web ¨Cbased image delivery. The wavelet transform, which is part of the new JPEG 2000 standard, claims to minimize some of the visually distracting artifacts that can appear in JPEG images. For one thing, it uses much larger blocks- selectable, but typically1024 x 1024 pixels ¨C for compression, rather than the 8 X 8 pixel blocks used in the original JPEG method, which often produced visible boundaries. Fractal compression has also shown promise and claims to be able to enlarge images by inserting ¨Drealistic¡¬ detail beyond the resolution limit of the original. Each method is discussed in the thesis

    Embed[d]ed Zerotree Codec

    Get PDF
    This thesis discusses the findings of the final year project involving the VHDL (V= Very High Speed Integrated Circuit, Hardware Description Language) design and simulation of an EZT (Embedded Zero Tree) codec. The basis of image compression and the various image compression techniques that are available today have been explored. This provided a clear understanding of image compression as a whole. An in depth understanding of wavelet transform theory was vital to the understanding of the edge that this transform provides over other transforms for image compression. Both the mathematics of it and how it is implemented using sets of high pass and low pass filters have been studied and presented. At the heart of the EZT codec is the EZW (Embedded Zerotree Wavelet) algorithm, as this is the algorithm that has been implemented in the codec. This required a thorough study and understanding of the algorithm and the various terms used in it. A generic single processor codec capable of handling any size of zerotree coefficients of images was designed. Once the coding and decoding strategy of this single processor had been figured out, it was easily extended to a codec with three parallel processors. This parallel architecture uses the same coding and decoding methods as in the single processor except that each processor in the parallel processing now handles only a third of the coefficients, thus promising a much speedier codec as compared to the first one. Both designs were then translated into VHDL behavioral level codes. The codes were then simulated and the results were verified. Once the simulations were completed the next aim for the project, namely synthesizing the design, was embarked upon. Of the two logical parts of the encoder, only the significance map generator has been synthesized

    Biometric Applications Based on Multiresolution Analysis Tools

    Get PDF
    This dissertation is dedicated to the development of new algorithms for biometric applications based on multiresolution analysis tools. Biometric is a unique, measurable characteristic of a human being that can be used to automatically recognize an individual or verify an individual\u27s identity. Biometrics can measure physiological, behavioral, physical and chemical characteristics of an individual. Physiological characteristics are based on measurements derived from direct measurement of a part of human body, such as, face, fingerprint, iris, retina etc. We focussed our investigations to fingerprint and face recognition since these two biometric modalities are used in conjunction to obtain reliable identification by various border security and law enforcement agencies. We developed an efficient and robust human face recognition algorithm for potential law enforcement applications. A generic fingerprint compression algorithm based on state of the art multiresolution analysis tool to speed up data archiving and recognition was also proposed. Finally, we put forth a new fingerprint matching algorithm by generating an efficient set of fingerprint features to minimize false matches and improve identification accuracy. Face recognition algorithms were proposed based on curvelet transform using kernel based principal component analysis and bidirectional two-dimensional principal component analysis and numerous experiments were performed using popular human face databases. Significant improvements in recognition accuracy were achieved and the proposed methods drastically outperformed conventional face recognition systems that employed linear one-dimensional principal component analysis. Compression schemes based on wave atoms decomposition were proposed and major improvements in peak signal to noise ratio were obtained in comparison to Federal Bureau of Investigation\u27s wavelet scalar quantization scheme. Improved performance was more pronounced and distinct at higher compression ratios. Finally, a fingerprint matching algorithm based on wave atoms decomposition, bidirectional two dimensional principal component analysis and extreme learning machine was proposed and noteworthy improvements in accuracy were realized

    Image coding for digitized libraries

    Get PDF
    Ankara : The Department of Electrical and Electronics Engineering and the Institute of Engineering and Sciences of Bilkent Univ., 1998.Thesis (Ph.D.) -- Bilkent University, 1998.Includes bibliographical references leaves 104-113III this thesis, image coding methods for two basic image types are developed under a digitized library framework. The two image types are gray tone or color images, and binary textual images, which are the digitized image versions of text documents. The grciy tone images are encoded using an adaptive subband decomposition followed by zerotree quantizers. The adaptive sub- l)and decomposition filter bank adaptively updates the filter bank coefficients in which the values of one of the subbands is predicted from the other sub- band. It is observed that the adaptive subband decomposition performs better than a regulcir subband decomposition with a fixed filter bank in terms of compression. For the binary textual images, a compression algorithm using binary subband decomposition followed by a textual image compression (TIC) method that exploits the redundancy in repeating characters is developed. The binary subband decomposition yields binary sub-images, and the TIC method is applied to the low band sub-image. Obtaining binary sub-images improves compression results as well as pattern matching time of the TIC method. Simulation results for both adaptive subband decomposition and multiresolution TIC methods indicate improvements over the methods described in the literature.Gerek, Ömer NezihPh.D

    A computational model of visual attention.

    Get PDF
    Visual attention is a process by which the Human Visual System (HVS) selects most important information from a scene. Visual attention models are computational or mathematical models developed to predict this information. The performance of the state-of-the-art visual attention models is limited in terms of prediction accuracy and computational complexity. In spite of significant amount of active research in this area, modelling visual attention is still an open research challenge. This thesis proposes a novel computational model of visual attention that achieves higher prediction accuracy with low computational complexity. A new bottom-up visual attention model based on in-focus regions is proposed. To develop the model, an image dataset is created by capturing images with in-focus and out-of-focus regions. The Discrete Cosine Transform (DCT) spectrum of these images is investigated qualitatively and quantitatively to discover the key frequency coefficients that correspond to the in-focus regions. The model detects these key coefficients by formulating a novel relation between the in-focus and out-of-focus regions in the frequency domain. These frequency coefficients are used to detect the salient in-focus regions. The simulation results show that this attention model achieves good prediction accuracy with low complexity. The prediction accuracy of the proposed in-focus visual attention model is further improved by incorporating sensitivity of the HVS towards the image centre and the human faces. Moreover, the computational complexity is further reduced by using Integer Cosine Transform (ICT). The model is parameter tuned using the hill climbing approach to optimise the accuracy. The performance has been analysed qualitatively and quantitatively using two large image datasets with eye tracking fixation ground truth. The results show that the model achieves higher prediction accuracy with a lower computational complexity compared to the state-of-the-art visual attention models. The proposed model is useful in predicting human fixations in computationally constrained environments. Mainly it is useful in applications such as perceptual video coding, image quality assessment, object recognition and image segmentation
    corecore