148 research outputs found
Progressive Lossless Image Compression Using Image Decomposition and Context Quantization
Lossless image compression has many applications, for example, in medical imaging, space photograph and film industry. In this thesis, we propose an efficient lossless image compression scheme for both binary images and gray-scale images. The scheme first decomposes images into a set of progressively refined binary sequences and then uses the context-based, adaptive arithmetic coding algorithm to encode these sequences. In order to deal with the context dilution problem in arithmetic coding, we propose a Lloyd-like iterative algorithm to quantize contexts. Fixing the set of input contexts and the number of quantized contexts, our context quantization algorithm iteratively finds the optimum context mapping in the sense of minimizing the compression rate. Experimental results show that by combining image decomposition and context quantization, our scheme can achieve competitive lossless compression performance compared to the JBIG algorithm for binary images, and the CALIC algorithm for gray-scale images. In contrast to CALIC, our scheme provides the additional feature of allowing progressive transmission of gray-scale images, which is very appealing in applications such as web browsing
Improvements to the color quantization process
The presentation of color images on devices with limited color capabilities requires a reduction in the number of colors contained in the images. Color image quantization is the process of reducing the number of colors used in an image while maintaining its appearance as much as possible. This reduction is performed using a color image quantization algorithm. The quantization algorithm attempts to select k colors that best represent the contents of the image. The original image is then recolored using the representative colors. to improve the resulting image, a dithering process can be used in place of the recoloring.;This dissertation deals with several areas of the color image quantization process. The main objective, however, is new or improved algorithms for the production of images with a better visual quality than those produced by existing algorithms while maintaining approximately the same running time. First, a new algorithm is developed for the selection of the representative color set. The results produced by the new algorithm are better both visually and quantitatively when compared to existing algorithms. Second, a new nearest-neighbor search algorithm that is based on the Locally Sorted Search algorithm is developed to reduce the time required to map the input colors to a representative color. Finally, two modifications are made to the error-diffusion dithering technique that improve the execution time. These modifications include the use of a two-weight matrix for the distribution of the error values and the presentation of a method to parallelize the error-diffusion technique. Furthermore, the analytical results of several experiments are provided to show the effectiveness each of these additions and improvements
Images on the Move
In contemporary society, digital images have become increasingly mobile. They are networked, shared on social media, and circulated across small and portable screens. Accordingly, the discourses of spreadability and circulation have come to supersede the focus on production, indexicality, and manipulability, which had dominated early conceptions of digital photography and film. However, the mobility of images is neither technologically nor conceptually limited to the realm of the digital. The edited volume re-examines the historical, aesthetical, and theoretical relevance of image mobility. The contributors provide a materialist account of images on the move - ranging from wired photography to postcards to streaming media
Images on the Move: Materiality - Networks - Formats
In contemporary society, digital images have become increasingly mobile. They are networked, shared on social media, and circulated across small and portable screens. Accordingly, the discourses of spreadability and circulation have come to supersede the focus on production, indexicality, and manipulability, which had dominated early conceptions of digital photography and film. However, the mobility of images is neither technologically nor conceptually limited to the realm of the digital. The edited volume re-examines the historical, aesthetical, and theoretical relevance of image mobility. The contributors provide a materialist account of images on the move - ranging from wired photography to postcards to streaming media
Images on the Move
In contemporary society, digital images have become increasingly mobile. They are networked, shared on social media, and circulated across small and portable screens. Accordingly, the discourses of spreadability and circulation have come to supersede the focus on production, indexicality, and manipulability, which had dominated early conceptions of digital photography and film. However, the mobility of images is neither technologically nor conceptually limited to the realm of the digital. The edited volume re-examines the historical, aesthetical, and theoretical relevance of image mobility. The contributors provide a materialist account of images on the move - ranging from wired photography to postcards to streaming media
Investigation of the effects of image compression on the geometric quality of digital protogrammetric imagery
We are living in a decade, where the use of digital images is becoming increasingly important. Photographs are now converted into digital form, and direct acquisition of digital images is becoming increasing important as sensors and associated electronics. Unlike images in analogue form, digital representation of images allows visual information to· be easily manipulated in useful ways. One practical problem of the digital image representation is that, it requires a very large number of bits and hence one encounters a fairly large volume of data in a digital production environment if they are stored uncompressed on the disk. With the rapid advances in sensor technology and digital electronics, the number of bits grow larger in softcopy photogrammetry, remote sensing and multimedia GIS. As a result, it is desirable to find efficient representation for digital images in order to reduce the memory required for storage, improve the data access rate from storage devices, and reduce the time required for transfer across communication channels. The component of digital image processing that deals with this problem is called image compression. Image compression is a necessity for the utilisation of large digital images in softcopy photogrammetry, remote sensing, and multimedia GIS. Numerous image Compression standards exist today with the common goal of reducing the number of bits needed to store images, and to facilitate the interchange of compressed image data between various devices and applications. JPEG image compression standard is one alternative for carrying out the image compression task. This standard was formed under the auspices ISO and CCITT for the purpose of developing an international standard for the compression and decompression of continuous-tone, still-frame, monochrome and colour images. The JPEG standard algorithm &Us into three general categories: the baseline sequential process that provides a simple and efficient algorithm for most image coding applications, the extended DCT-based process that allows the baseline system to satisfy a broader range of applications, and an independent lossless process for application demanding that type of compression. This thesis experimentally investigates the geometric degradations resulting from lossy JPEG compression on photogrammetric imagery at various levels of quality factors. The effects and the suitability of JPEG lossy image compression on industrial photogrammetric imagery are investigated. Examples are drawn from the extraction of targets in close-range photogrammetric imagery. In the experiments, the JPEG was used to compress and decompress a set of test images. The algorithm has been tested on digital images containing various levels of entropy (a measure of information content of an image) with different image capture capabilities. Residual data was obtained by taking the pixel-by-pixel difference between the original data and the reconstructed data. The image quality measure, root mean square (rms) error of the residual was used as a quality measure to judge the quality of images produced by JPEG(DCT-based) image compression technique. Two techniques, TIFF (IZW) compression and JPEG(DCT-based) compression are compared with respect to compression ratios achieved. JPEG(DCT-based) yields better compression ratios, and it seems to be a good choice for image compression. Further in the investigation, it is found out that, for grey-scale images, the best compression ratios were obtained when the quality factors between 60 and 90 were used (i.e., at a compression ratio of 1:10 to 1:20). At these quality factors the reconstructed data has virtually no degradation in the visual and geometric quality for the application at hand. Recently, many fast and efficient image file formats have also been developed to store, organise and display images in an efficient way. Almost every image file format incorporates some kind of compression method to manage data within common place networks and storage devices. The current major file formats used in softcopy photogrammetry, remote sensing and · multimedia GIS. were also investigated. It was also found out that the choice of a particular image file format for a given application generally involves several interdependent considerations including quality; flexibility; computation; storage, or transmission. The suitability of a file format for a given purpose is · best determined by knowing its original purpose. Some of these are widely used (e.g., TIFF, JPEG) and serve as exchange formats. Others are adapted to the needs of particular applications or particular operating systems
Whole Blood Mitochondrial DNA Depletion in Human Immunodeficiency Virus-Infected Children
Background: Nucleoside reverse transcriptase inhibitors (NRTIs) interfere with mitochondrial DNA polymerase gamma causing significant toxic effects, including fatal lactic acidosis. Little is known about mitochondrial DNA (mtDNA) in human immunodeficiency virus (HIV) infected children who face a lifetime exposure to these agents. We performed a cross sectional observation of mtDNA levels in whole blood in a pediatric population to ascertain the relationship between mtDNA, NRTI regimens and parameters of HIV-infection severity. Methods: Whole blood mt:nDNA ratios were determined by real-time PCR in three groups: 27 presumed HIV-negative, 89 HIV-infected, NRTI-treated and 62 HIV-infected treatment-naive children. Multivariate analysis was used to identify variables independently associated with mtDNA depletion. Results: Mean mt:nDNA ratios were lower (P < 0.001) at 77% of control in the HIVinfected antiretroviral treatment (ART) Naïve group and 73% of control in the ART group, but not different between the two HIV-infected groups. Mt:nDNA ratios were negatively associated with age (P = 0.029), HIV status (P < 0.0001) and Log10 of the HIV-1 viral load (P = 0.035) and positively associated with CD4 % (p = 0.032). A 6 stavudine vs zidovudine based regimen was associated with lower but not significant levels of mtDNA (P = 0.1). Conclusions: Depletion of whole blood mtDNA in children is associated independently with HIV-infection and markers of HIV infection severity, and does not improve with either stavudine or zidovudine based ART despite virological control, suggesting that these agents also deplete mtDNA
- …