526 research outputs found
Recovering Sign Bits of DCT Coefficients in Digital Images as an Optimization Problem
Recovering unknown, missing, damaged, distorted or lost information in DCT
coefficients is a common task in multiple applications of digital image
processing, including image compression, selective image encryption, and image
communications. This paper investigates recovery of a special type of
information in DCT coefficients of digital images: sign bits. This problem can
be modelled as a mixed integer linear programming (MILP) problem, which is
NP-hard in general. To efficiently solve the problem, we propose two
approximation methods: 1) a relaxation-based method that convert the MILP
problem to a linear programming (LP) problem; 2) a divide-and-conquer method
which splits the target image into sufficiently small regions, each of which
can be more efficiently solved as an MILP problem, and then conducts a global
optimization phase as a smaller MILP problem or an LP problem to maximize
smoothness across different regions. To the best of our knowledge, we are the
first who considered how to use global optimization to recover sign bits of DCT
coefficients. We considered how the proposed methods can be applied to
JPEG-encoded images and conducted extensive experiments to validate the
performances of our proposed methods. The experimental results showed that the
proposed methods worked well, especially when the number of unknown sign bits
per DCT block is not too large. Compared with other existing methods, which are
all based on simple error-concealment strategies, our proposed methods
outperformed them with a substantial margin, both according to objective
quality metrics (PSNR and SSIM) and also our subjective evaluation. Our work
has a number of profound implications, e.g., more sign bits can be discarded to
develop more efficient image compression methods, and image encryption methods
based on sign bit encryption can be less secure than we previously understood.Comment: 13 pages, 8 figure
Literature Study On Cloud Based Healthcare File Protection Algorithms
There is a huge development in Computers and Cloud computing technology, the trend in recent years is to outsource information storage on Cloud-based services. The cloud provides large storage space. Cloud-based service providers such as Dropbox, Google Drive, are providing users with infinite and low-cost storage. In this project we aim at presenting a protection method through by encrypting and decrypting the files to provide enhanced level of protection. To encrypt the file that we upload in cloud, we make use of double encryption technique. The file is been encrypted twice one followed by the other using two algorithms. The order in which the algorithms are used is that, the file is first encrypted using AES algorithm, now this file will be in the encrypted format and this encrypted file is again encrypted using RSA algorithm. The corresponding keys are been generated during the execution of the algorithm. This is done in order to increase the security level. The various parameters that we have considered here are security level, speed, data confidentiality, data integrity and cipher text size. Our project is more efficient as it satisfies all the parameters whereas the conventional methods failed to do so. The Cloud we used is Dropbox to store the content of the file which is in the encrypted format using AES and RSA algorithms and corresponding key is generated which can be used to decrypt the file. While uploading the file the double encryption technique is been implemented
Literature Study on Data Protection for Cloud Storage
Many data security and privacy incidents are observed in today Cloud services. On the one hand, Cloud service providers deal with a large number of external attacks. In 2018, a total of 1.5 million Sing Health patients’ non-medical personal data were stolen from the health system in Singapore. On the other hand, Cloud service providers cannot be entirely trusted either. Personal data may be exploited in a malicious way such as in the Face book and Cambridge Analytical data scandal which affected 87 million users in 2018. Thus, it becomes increasingly important for end users to efficiently protect their data (texts, images, or videos) independently from Cloud service providers. In this paper, we aim at presenting a novel data protection scheme by combining fragmentation, encryption, and dispersion with high performance and enhanced level of protection as Literature study
Study and Implementation of Watermarking Algorithms
Water Making is the process of embedding data called a watermark into a multimedia object such that watermark can be detected or extracted later to make an assertion about the object. The object may be an audio, image or video. A copy of a digital image is identical to the original. This has in many instances, led to the use of digital content with malicious intent. One way to protect multimedia data against illegal recording and retransmission is to embed a signal, called digital signature or copyright label or watermark that authenticates the owner of the data. Data hiding, schemes to embed secondary data in digital media, have made considerable progress in recent years and attracted attention from both academia and industry. Techniques have been proposed for a variety of applications, including ownership protection, authentication and access control. Imperceptibility, robustness against moderate processing such as compression, and the ability to hide many bits are the basic but rat..
Natural Image Statistics for Digital Image Forensics
We describe a set of natural image statistics that are built upon two multi-scale image decompositions, the quadrature mirror filter pyramid decomposition and the local angular harmonic decomposition. These image statistics consist of first- and higher-order statistics that capture certain statistical regularities of natural images. We propose to apply these image statistics, together with classification techniques, to three problems in digital image forensics: (1) differentiating photographic images from computer-generated photorealistic images, (2) generic steganalysis; (3) rebroadcast image detection. We also apply these image statistics to the traditional art authentication for forgery detection and identification of artists in an art work. For each application we show the effectiveness of these image statistics and analyze their sensitivity and robustness
Recommended from our members
Secure digital documents using Steganography and QR Code
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University LondonWith the increasing use of the Internet several problems have arisen regarding the processing of electronic documents. These include content filtering, content retrieval/search. Moreover, document security has taken a centre stage including copyright protection, broadcast monitoring etc. There is an acute need of an effective tool which can find the identity, location and the time when the document was created so that it can be determined whether or not the contents of the document were tampered with after creation. Owing the sensitivity of the large amounts of data which is processed on a daily basis, verifying the authenticity and integrity of a document is more important now than it ever was. Unsurprisingly document authenticity verification has become the centre of attention in the world of research. Consequently, this research is concerned with creating a tool which deals with the above problem. This research proposes the use of a Quick Response Code as a message carrier for Text Key-print. The Text Key-print is a novel method which employs the basic element of the language (i.e. Characters of the alphabet) in order to achieve authenticity of electronic documents through the transformation of its physical structure into a logical structured relationship. The resultant dimensional matrix is then converted into a binary stream and encapsulated with a serial number or URL inside a Quick response Code (QR code) to form a digital fingerprint mark. For hiding a QR code, two image steganography techniques were developed based upon the spatial and the transform domains. In the spatial domain, three methods were proposed and implemented based on the least significant bit insertion technique and the use of pseudorandom number generator to scatter the message into a set of arbitrary pixels. These methods utilise the three colour channels in the images based on the RGB model based in order to embed one, two or three bits per the eight bit channel which results in three different hiding capacities. The second technique is an adaptive approach in transforming domain where a threshold value is calculated under a predefined location for embedding in order to identify the embedding strength of the embedding technique. The quality of the generated stego images was evaluated using both objective (PSNR) and Subjective (DSCQS) methods to ensure the reliability of our proposed methods. The experimental results revealed that PSNR is not a strong indicator of the perceived stego image quality, but not a bad interpreter also of the actual quality of stego images. Since the visual difference between the cover and the stego image must be absolutely imperceptible to the human visual system, it was logically convenient to ask human observers with different qualifications and experience in the field of image processing to evaluate the perceived quality of the cover and the stego image. Thus, the subjective responses were analysed using statistical measurements to describe the distribution of the scores given by the assessors. Thus, the proposed scheme presents an alternative approach to protect digital documents rather than the traditional techniques of digital signature and watermarking
Blind Image Watermark Detection Algorithm based on Discrete Shearlet Transform Using Statistical Decision Theory
Blind watermarking targets the challenging recovery of the watermark when the host is not available during the detection stage.This paper proposes Discrete Shearlet Transform as a new embedding domain for blind image watermarking. Our novel DST blind watermark detection system uses a nonadditive scheme based on the statistical decision theory. It first computes the probability density function (PDF) of the DST coefficients modelled as a Laplacian distribution. The resulting likelihood ratio is compared with a decision threshold calculated using Neyman-Pearson criterion to minimise the missed detection subject to a fixed false alarm probability. Our method is evaluated in terms of imperceptibility, robustness and payload against different attacks (Gaussian noise, Blurring, Cropping, Compression and Rotation) using 30 standard grayscale images covering different characteristics (smooth, more complex with a lot of edges and high detail textured regions). The proposed method shows greater windowing flexibility with more sensitive to directional and anisotropic features when compared against Discrete Wavelet and Contourlets
- …