2,356 research outputs found

    On the Application of Dictionary Learning to Image Compression

    Get PDF
    Signal models are a cornerstone of contemporary signal and image-processing methodology. In this chapter, a particular signal modelling method, called synthesis sparse representation, is studied which has been proven to be effective for many signals, such as natural images, and successfully used in a wide range of applications. In this kind of signal modelling, the signal is represented with respect to dictionary. The dictionary choice plays an important role on the success of the entire model. One main discipline of dictionary designing is based on a machine learning methodology which provides a simple and expressive structure for designing adaptable and efficient dictionaries. This chapter focuses on direct application of the sparse representation, i.e. image compression. Two image codec based on adaptive sparse representation over a trained dictionary are introduced. Experimental results show that the presented methods outperform the existing image coding standards, such as JPEG and JPEG2000

    Context-Patch Face Hallucination Based on Thresholding Locality-Constrained Representation and Reproducing Learning

    Get PDF
    Face hallucination is a technique that reconstruct high-resolution (HR) faces from low-resolution (LR) faces, by using the prior knowledge learned from HR/LR face pairs. Most state-of-the-arts leverage position-patch prior knowledge of human face to estimate the optimal representation coefficients for each image patch. However, they focus only the position information and usually ignore the context information of image patch. In addition, when they are confronted with misalignment or the Small Sample Size (SSS) problem, the hallucination performance is very poor. To this end, this study incorporates the contextual information of image patch and proposes a powerful and efficient context-patch based face hallucination approach, namely Thresholding Locality-constrained Representation and Reproducing learning (TLcR-RL). Under the context-patch based framework, we advance a thresholding based representation method to enhance the reconstruction accuracy and reduce the computational complexity. To further improve the performance of the proposed algorithm, we propose a promotion strategy called reproducing learning. By adding the estimated HR face to the training set, which can simulates the case that the HR version of the input LR face is present in the training set, thus iteratively enhancing the final hallucination result. Experiments demonstrate that the proposed TLcR-RL method achieves a substantial increase in the hallucinated results, both subjectively and objectively. Additionally, the proposed framework is more robust to face misalignment and the SSS problem, and its hallucinated HR face is still very good when the LR test face is from the real-world. The MATLAB source code is available at https://github.com/junjun-jiang/TLcR-RL

    An Inference about Interference: A Surprising Application of Existing International Law to Inhibit Anti-Satellite Weapons

    Get PDF
    This article presents a thesis that most readers will find surprising, in an effort to develop a novel, simultaneous solution to three urgent, complex problems related to outer space. The three problems are: a) the technical fact that debris in outer space (the accumulated orbital junk produced by decades of space activities) has grown to present a serious hazard to safe and effective exploration and exploitation of space; b) the strategic fact that many countries (notably the United States, China and Russia, but others, too) continue to demonstrate a misguided interest in pursuing anti-satellite weapons, which can jeopardize the security of space; and c) the political fact that attempts to provide additional legal regulation of outer space (via new bilateral or multilateral international agreements) have failed, with little prospect for prompt conclusion of meaningful new accords. The proposed solution is to adapt existing international law in an unforeseen way. Specifically, numerous current and historical arms control treaties provide for verification of parties’ compliance via “national technical means” (NTM) of verification, which prominently include satellite-based sensory and communications systems. These treaties routinely provide protection for those essential space assets by requiring parties to undertake “not to interfere” with NTM. The argument developed here is that additional tests in space of debris-creating anti-satellite weapons would already be illegal, even without the conclusion of any dedicated new treaty against further weaponization of space, because in the current crowded conditions of space, a new cloud of orbital debris would, sooner or later, impermissibly interfere with NTM satellites. If sustained, this thesis can provide a new rationale for opposition to the development, testing, and use of anti-satellite weapons. It a legal reinforcement for the political instincts to avoid activities that further undercut the optimal usability of outer space, and it demonstrates how creative re-interpretation of existing legal provisions can promote the advancement of the rule of international law, even in circumstances where the articulation of new treaties is blocked

    Enhanced information extraction in the multi-energy x-ray tomography for security

    Full text link
    Thesis (Ph.D.)--Boston UniversityX-ray Computed Tomography (CT) is an effective nondestructive technology widely used for medical diagnosis and security. In CT, three-dimensional images of the interior of an object are generated based on its X-ray attenuation. Conventional CT is performed with a single energy spectrum and materials can only be differentiated based on an averaged measure of the attenuation. Multi-Energy CT (MECT) methods have been developed to provide more information about the chemical composition of the scanned material using multiple energy-selective measurements of the attenuation. Existing literature on MECT is mostly focused on differentiation between body tissues and other medical applications. The problems in security are more challenging due to the larger range of materials and threats which may be found. Objects may appear in high clutter and in different forms of concealment. Thus, the information extracted by the medical domain methods may not be optimal for detection of explosives and improved performance is desired. In this dissertation, learning and adaptive model-based methods are developed to address the challenges of multi-energy material discrimination for security. First, the fundamental information contained in the X-ray attenuation versus energy curves of materials is studied. For this purpose, a database of these curves for a set of explosive and non-explosive compounds was created. The dimensionality and span of the curves is estimated and their space is shown to be larger than two-dimensional, contrary to what is typically assumed. In addition, optimized feature selection methods are developed and applied to the curves and it is demonstrated that detection performance may be improved by using more than two features and when using features different than the standard photoelectric and Compton coefficients. Second, several MECT reconstruction methods are studied and compared. This includes a new structure-preserving inversion technique which can mitigate metal artifacts and provide precise object localization in the estimated parameter images. Finally, a learning-based MECT framework for joint material classification and segmentation is developed, which can produce accurate material labels in the presence of metal and clutter. The methods are tested on simulated and real multi-energy data and it is shown that they outperform previously published MECT techniques

    Mathematical Approaches for Image Enhancement Problems

    Get PDF
    This thesis develops novel techniques that can solve some image enhancement problems using theoretically and technically proven and very useful mathematical tools to image processing such as wavelet transforms, partial differential equations, and variational models. Three subtopics are mainly covered. First, color image denoising framework is introduced to achieve high quality denoising results by considering correlations between color components while existing denoising approaches can be plugged in flexibly. Second, a new and efficient framework for image contrast and color enhancement in the compressed wavelet domain is proposed. The proposed approach is capable of enhancing both global and local contrast and brightness as well as preserving color consistency. The framework does not require inverse transform for image enhancement since linear scale factors are directly applied to both scaling and wavelet coefficients in the compressed domain, which results in high computational efficiency. Also contaminated noise in the image can be efficiently reduced by introducing wavelet shrinkage terms adaptively in different scales. The proposed method is able to enhance a wavelet-coded image computationally efficiently with high image quality and less noise or other artifact. The experimental results show that the proposed method produces encouraging results both visually and numerically compared to some existing approaches. Finally, image inpainting problem is discussed. Literature review, psychological analysis, and challenges on image inpainting problem and related topics are described. An inpainting algorithm using energy minimization and texture mapping is proposed. Mumford-Shah energy minimization model detects and preserves edges in the inpainting domain by detecting both the main structure and the detailed edges. This approach utilizes faster hierarchical level set method and guarantees convergence independent of initial conditions. The estimated segmentation results in the inpainting domain are stored in segmentation map, which is referred by a texture mapping algorithm for filling textured regions. We also propose an inpainting algorithm using wavelet transform that can expect better global structure estimation of the unknown region in addition to shape and texture properties since wavelet transforms have been used for various image analysis problems due to its nice multi-resolution properties and decoupling characteristics

    Video modeling via implicit motion representations

    Get PDF
    Video modeling refers to the development of analytical representations for explaining the intensity distribution in video signals. Based on the analytical representation, we can develop algorithms for accomplishing particular video-related tasks. Therefore video modeling provides us a foundation to bridge video data and related-tasks. Although there are many video models proposed in the past decades, the rise of new applications calls for more efficient and accurate video modeling approaches.;Most existing video modeling approaches are based on explicit motion representations, where motion information is explicitly expressed by correspondence-based representations (i.e., motion velocity or displacement). Although it is conceptually simple, the limitations of those representations and the suboptimum of motion estimation techniques can degrade such video modeling approaches, especially for handling complex motion or non-ideal observation video data. In this thesis, we propose to investigate video modeling without explicit motion representation. Motion information is implicitly embedded into the spatio-temporal dependency among pixels or patches instead of being explicitly described by motion vectors.;Firstly, we propose a parametric model based on a spatio-temporal adaptive localized learning (STALL). We formulate video modeling as a linear regression problem, in which motion information is embedded within the regression coefficients. The coefficients are adaptively learned within a local space-time window based on LMMSE criterion. Incorporating a spatio-temporal resampling and a Bayesian fusion scheme, we can enhance the modeling capability of STALL on more general videos. Under the framework of STALL, we can develop video processing algorithms for a variety of applications by adjusting model parameters (i.e., the size and topology of model support and training window). We apply STALL on three video processing problems. The simulation results show that motion information can be efficiently exploited by our implicit motion representation and the resampling and fusion do help to enhance the modeling capability of STALL.;Secondly, we propose a nonparametric video modeling approach, which is not dependent on explicit motion estimation. Assuming the video sequence is composed of many overlapping space-time patches, we propose to embed motion-related information into the relationships among video patches and develop a generic sparsity-based prior for typical video sequences. First, we extend block matching to more general kNN-based patch clustering, which provides an implicit and distributed representation for motion information. We propose to enforce the sparsity constraint on a higher-dimensional data array signal, which is generated by packing the patches in the similar patch set. Then we solve the inference problem by updating the kNN array and the wanted signal iteratively. Finally, we present a Bayesian fusion approach to fuse multiple-hypothesis inferences. Simulation results in video error concealment, denoising, and deartifacting are reported to demonstrate its modeling capability.;Finally, we summarize the proposed two video modeling approaches. We also point out the perspectives of implicit motion representations in applications ranging from low to high level problems

    Robust density modelling using the student's t-distribution for human action recognition

    Full text link
    The extraction of human features from videos is often inaccurate and prone to outliers. Such outliers can severely affect density modelling when the Gaussian distribution is used as the model since it is highly sensitive to outliers. The Gaussian distribution is also often used as base component of graphical models for recognising human actions in the videos (hidden Markov model and others) and the presence of outliers can significantly affect the recognition accuracy. In contrast, the Student's t-distribution is more robust to outliers and can be exploited to improve the recognition rate in the presence of abnormal data. In this paper, we present an HMM which uses mixtures of t-distributions as observation probabilities and show how experiments over two well-known datasets (Weizmann, MuHAVi) reported a remarkable improvement in classification accuracy. © 2011 IEEE
    • …
    corecore