264 research outputs found
Geodesics on the manifold of multivariate generalized Gaussian distributions with an application to multicomponent texture discrimination
We consider the Rao geodesic distance (GD) based on the Fisher information as a similarity measure on the manifold of zero-mean multivariate generalized Gaussian distributions (MGGD). The MGGD is shown to be an adequate model for the heavy-tailed wavelet statistics in multicomponent images, such as color or multispectral images. We discuss the estimation of MGGD parameters using various methods. We apply the GD between MGGDs to color texture discrimination in several classification experiments, taking into account the correlation structure between the spectral bands in the wavelet domain. We compare the performance, both in terms of texture discrimination capability and computational load, of the GD and the Kullback-Leibler divergence (KLD). Likewise, both uni- and multivariate generalized Gaussian models are evaluated, characterized by a fixed or a variable shape parameter. The modeling of the interband correlation significantly improves classification efficiency, while the GD is shown to consistently outperform the KLD as a similarity measure
Wavelet based similarity measurement algorithm for seafloor morphology
Thesis (S.M. in Naval Architecture and Marine Engineering and S.M. in Mechanical Engineering)--Massachusetts Institute of Technology, Dept. of Mechanical Engineering, 2006.Includes bibliographical references (leaves 71-73).The recent expansion of systematic seafloor exploration programs such as geophysical research, seafloor mapping, search and survey, resource assessment and other scientific, commercial and military applications has created a need for rapid and robust methods of processing seafloor imagery. Given the existence of a large library of seafloor images, a fast automated image classifier algorithm is needed to determine changes in seabed morphology over time. The focus of this work is the development of a robust Similarity Measurement (SM) algorithm to address the above problem. Our work uses a side-scan sonar image library for experimentation and testing. Variations of an underwater vehicle's height above the sea floor and of its pitch and roll angles cause distortion in the data obtained, such that transformations to align the data should include rotation, translation, anisotropic scaling and skew. In order to deal with these problems, we propose to use the Wavelet transform for similarity detection. Wavelets have been widely used during the last three decades in image processing. Since the Wavelet transform allows a multi-resolution decomposition, it is easier to identify the similarities between two images by examining the energy distribution at each decomposition level.(cont.) The energy distribution in the frequency domain at the output of the high pass and low pass filter banks identifies the texture discrimination. Our approach uses a statistical framework, involving fitting the Wavelet coefficients into a generalized Gaussian density distribution. The next step involves use of the Kullback-Leibner entropy metric to measure the distance between Wavelet coefficient distributions. To select the top N most likely matching images, the database images are ranked based on the minimum Kullback-Leibner distance. The statistical approach is effective in eliminating rotation, mis-registration and skew problems by working in the Wavelet domain. It's recommended that further work focuses on choosing the best Wavelet packet to increase the robustness of the algorithm developed in this thesis.by Ilkay Darilmaz.S.M.in Naval Architecture and Marine Engineering and S.M.in Mechanical Engineerin
Dual-tree Complex Wavelet Transform based Local Binary Pattern Weighted Histogram Method for Palmprint Recognition
In the paper, we improve the Local Binary Pattern Histogram (LBPH) approach and combine it with Dual-Tree Complex Wavelet Transform (DT-CWT) to propose a Dual-Tree Complex Wavelet Transform based Local Binary Pattern Weighted Histogram (DT-CWT based LBPWH) method for palmprint representation and recognition. The approximate shift invariant property of the DT-CWT and its good directional selectively in 2D make it a very appealing choice for palmprint representation. LBPH is a powerful texture description method, which considers both shape and texture information to represent an image. To enhance the representation capability of LBPH, a weight set is computed and assigned to the finial feature histogram. Here we needn't construct a palmprint model by a train sample set, which is not like some methods based on subspace discriminant analysis or statistical learning. In the approach, a palmprint image is first decomposed into multiple subbands by using DT-CWT. After that, each subband in complex wavelet domain is divided into non-overlapping sub-regions. Then LBPHs are extracted from each sub-region in each subband, and lastly, all of LBPHs are weighted and concatenated into a single feature histogram to effectively represent the palmprint image. A Chi square distance is used to measure the similarity of different feature histograms and the finial recognition is performed by the nearest neighborhood classifier. A group of optimal parameters is chosen by 20 verification tests on our palmprint database. In addition, the recognition results on our palmprint database and the database from the Hong Kong Polytechnic University show the proposed method outperforms other methods
A novel coarse-to-fine remote sensing image retrieval system in JPEG-2000 compressed domain
Copyright 2018 Society of Photo‑Optical Instrumentation Engineers (SPIE). One print or electronic copy may be made for personal use only. Systematic reproduction and distribution, duplication of any material in this publication for a fee or for commercial purposes, and modification of the contents of the publication are prohibited.This paper presents a novel content-based image search and retrieval (CBIR) system that achieves coarse to fine remote sensing (RS) image description and retrieval in JPEG 2000 compressed domain. The proposed system initially: i) decodes the code-streams associated to the coarse (i.e., the lowest) wavelet resolution, and ii) discards the most irrelevant images to the query image that are selected based on the similarities estimated among the coarse resolution features of the query image and those of the archive images. Then, the code-streams associated to the sub-sequent resolution of the remaining images in the archive are decoded and the most irrelevant images are selected by considering the features associated to both resolutions. This is achieved by estimating the similarities between the query image and remaining images by giving higher weights to the features associated to the finer resolution while assigning lower weights to those related to the coarse resolution. To this end, the pyramid match kernel similarity measure is exploited. These processes are iterated until the code-streams associated to the highest wavelet resolution are decoded only for a very small set of images. By this way, the proposed system exploits a multiresolution and hierarchical feature space and accomplish an adaptive RS CBIR with significantly reduced retrieval time. Experimental results obtained on an archive of aerial images confirm the effectiveness of the proposed system in terms of retrieval accuracy and time when compared to the standard CBIR systems
Perceptual Image Similarity Metrics and Applications.
This dissertation presents research in perceptual image similarity metrics and applications, e.g., content-based image retrieval, perceptual image compression, image similarity assessment and texture analysis.
The first part aims to design texture similarity metrics consistent with human perception. A new family of statistical texture similarity features, called Local Radius Index (LRI), and corresponding similarity metrics are proposed. Compared to state-of-the-art metrics in the STSIM family, LRI-based metrics achieve better texture retrieval performance with much less computation. When applied to the recently developed perceptual image coder, Matched Texture Coding (MTC), they enable similar performance while significantly accelerating encoding. Additionally, in photographic paper classification, LRI-based metrics also outperform pre-existing metrics. To fulfill the needs of texture classification and other applications, a rotation-invariant version of LRI, called Rotation-Invariant Local Radius Index (RI-LRI), is proposed. RI-LRI is also grayscale and illuminance insensitive. The corresponding similarity metric achieves texture classification accuracy comparable to state-of-the-art metrics. Moreover, its much lower dimensional feature vector requires substantially less computation and storage than other state-of-the-art texture features.
The second part of the dissertation focuses on bilevel images, which are images whose pixels are either black or white. The contributions include new objective similarity metrics intended to quantify similarity consistent with human perception, and a subjective experiment to obtain ground truth for judging the performance of objective metrics. Several similarity metrics are proposed that outperform existing ones in the sense of attaining significantly higher Pearson and Spearman-rank correlations with the ground truth. The new metrics include Adjusted Percentage Error, Bilevel Gradient Histogram, Connected Components Comparison and combinations of such.
Another portion of the dissertation focuses on the aforementioned MTC, which is a block-based image coder that uses texture similarity metrics to decide if blocks of the image can be encoded by pointing to perceptually similar ones in the already coded region. The key to its success is an effective texture similarity metric, such as an LRI-based metric, and an effective search strategy. Compared to traditional image compression algorithms, e.g., JPEG, MTC achieves similar coding rate with higher reconstruction quality. And the advantage of MTC becomes larger as coding rate decreases.PhDElectrical Engineering: SystemsUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttp://deepblue.lib.umich.edu/bitstream/2027.42/113586/1/yhzhai_1.pd
Wavelet-based texture retrieval using generalized Gaussian density and Kullback-Leibler distance
We present a statistical view of the texture retrieval problem by combining the two related tasks, namely feature extraction (FE) and similarity measurement (SM), into a joint modeling and classification scheme. We show that using a con- sistent estimator of texture model parameters for the FE step followed by computing the Kullback–Leibler distance (KLD) between estimated models for the SM step is asymptotically optimal in term of retrieval error probability. The statistical scheme leads to a new wavelet-based texture retrieval method that is based on the accurate modeling of the marginal distribution of wavelet coefficients using generalized Gaussian density (GGD) and on the existence a closed form for the KLD between GGDs. The proposed method provides greater accuracy and flexibility in capturing texture information, while its simplified form has a close resemblance with the existing methods which uses energy distribution in the frequency domain to identify textures. Ex- perimental results on a database of 640 texture images indicate that the new method significantly improves retrieval rates, e.g., from 65% to 77%, compared with traditional approaches, while it retains comparable levels of computational complexity
Image-Based Query by Example Using MPEG-7 Visual Descriptors
This project presents the design and implementation of a Content-Based Image Retrieval
(CBIR) system where queries are formulated by visual examples through a
graphical interface. Visual descriptors and similarity measures implemented in this
work followed mainly those defined in the MPEG-7 standard although, when necessary,
extensions are proposed. Despite the fact that this is an image-based system, all
the proposed descriptors have been implemented for both image and region queries,
allowing the future system upgrade to support region-based queries. This way, even a
contour shape descriptor has been developed, which has no sense for the whole image.
The system has been assessed on different benchmark databases; namely, MPEG-7
Common Color Dataset, and Corel Dataset. The evaluation has been performed for
isolated descriptors as well as for combinations of them. The strategy studied in this
work to gather the information obtained from the whole set of computed descriptors
is weighting the rank list for each isolated descriptor
Recommended from our members
Video content analysis for automated detection and tracking of humans in CCTV surveillance applications
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.The problems of achieving high detection rate with low false alarm rate for human detection and tracking in video sequence, performance scalability, and improving response time are addressed in this thesis. The underlying causes are the effect of scene complexity, human-to-human interactions, scale changes, and scene background-human interactions. A two-stage processing solution, namely, human detection, and human tracking with two novel pattern classifiers is presented. Scale independent human detection is achieved by processing in the wavelet domain using square wavelet features. These features used to characterise human silhouettes at different scales are similar to rectangular features used in [Viola 2001]. At the detection stage two detectors are combined to improve detection rate. The first detector is based on shape-outline of humans extracted from the scene using a reduced complexity outline extraction algorithm. A Shape mismatch measure is used to differentiate between the human and the background class. The second detector uses rectangular features as primitives for silhouette description in the wavelet domain. The marginal distribution of features collocated at a particular position on a candidate human (a patch of the image) is used to describe statistically the silhouette. Two similarity measures are computed between a candidate human and the model histograms of human and non human classes. The similarity measure is used to discriminate between the human and the non human class. At the tracking stage, a tracker based on joint probabilistic data association filter (JPDAF) for data association, and motion correspondence is presented. Track clustering is used to reduce hypothesis enumeration complexity. Towards improving response time with increase in frame dimension, scene complexity, and number of channels; a scalable algorithmic architecture and operating accuracy prediction technique is presented. A scheduling strategy for improving the response time and throughput by parallel processing is also presented
Visual Quality Assessment and Blur Detection Based on the Transform of Gradient Magnitudes
abstract: Digital imaging and image processing technologies have revolutionized the way in which
we capture, store, receive, view, utilize, and share images. In image-based applications,
through different processing stages (e.g., acquisition, compression, and transmission), images
are subjected to different types of distortions which degrade their visual quality. Image
Quality Assessment (IQA) attempts to use computational models to automatically evaluate
and estimate the image quality in accordance with subjective evaluations. Moreover, with
the fast development of computer vision techniques, it is important in practice to extract
and understand the information contained in blurred images or regions.
The work in this dissertation focuses on reduced-reference visual quality assessment of
images and textures, as well as perceptual-based spatially-varying blur detection.
A training-free low-cost Reduced-Reference IQA (RRIQA) method is proposed. The
proposed method requires a very small number of reduced-reference (RR) features. Extensive
experiments performed on different benchmark databases demonstrate that the proposed
RRIQA method, delivers highly competitive performance as compared with the
state-of-the-art RRIQA models for both natural and texture images.
In the context of texture, the effect of texture granularity on the quality of synthesized
textures is studied. Moreover, two RR objective visual quality assessment methods that
quantify the perceived quality of synthesized textures are proposed. Performance evaluations
on two synthesized texture databases demonstrate that the proposed RR metrics outperforms
full-reference (FR), no-reference (NR), and RR state-of-the-art quality metrics in
predicting the perceived visual quality of the synthesized textures.
Last but not least, an effective approach to address the spatially-varying blur detection
problem from a single image without requiring any knowledge about the blur type, level,
or camera settings is proposed. The evaluations of the proposed approach on a diverse
sets of blurry images with different blur types, levels, and content demonstrate that the
proposed algorithm performs favorably against the state-of-the-art methods qualitatively
and quantitatively.Dissertation/ThesisDoctoral Dissertation Electrical Engineering 201
- …