123 research outputs found
Enhancement of Single and Composite Images Based on Contourlet Transform Approach
Image enhancement is an imperative step in almost every image processing algorithms.
Numerous image enhancement algorithms have been developed for gray scale images
despite their absence in many applications lately. This thesis proposes hew image
enhancement techniques of 8-bit single and composite digital color images. Recently, it
has become evident that wavelet transforms are not necessarily best suited for images.
Therefore, the enhancement approaches are based on a new 'true' two-dimensional
transform called contourlet transform. The proposed enhancement techniques discussed
in this thesis are developed based on the understanding of the working mechanisms of the
new multiresolution property of contourlet transform. This research also investigates the
effects of using different color space representations for color image enhancement
applications. Based on this investigation an optimal color space is selected for both single
image and composite image enhancement approaches. The objective evaluation steps
show that the new method of enhancement not only superior to the commonly used
transformation method (e.g. wavelet transform) but also to various spatial models (e.g.
histogram equalizations). The results found are encouraging and the enhancement
algorithms have proved to be more robust and reliable
Recommended from our members
Deep learning assisted MRI guided attenuation correction in PET
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University LondonPositron emission tomography (PET) is a unique imaging modality that provides physiological
and functional details of the tissue at the molecular level. However, the acquired PET images
have some limitations such as the attenuation. PET attenuation correction is an essential step to
obtain the full potential of PET quantification. With the wide use of hybrid PET/MR scanners,
magnetic resonance (MR) images are used to address the problem of PET attenuation correction.
The MR images segmentation is one simple and robust approach to create pseudo computed
tomography (CT) images, which are used to generate attenuation coefficient maps to correct the
PET attenuation. Recently, deep learning has been proposed and used as a promising technique
to efficiently perform MR and various medical images segmentation.
In this research work, deep learning guided segmentation approaches have been proposed
to enhance the bone class segmentation of MR brain images in order to generate accurate
pseudo-CT images. The first approach has introduced the combination of handcrafted features
with deep learning features to enrich the set of features. Multiresolution analysis techniques,
which generate multiscale and multidirectional coefficients of an image such as contourlet and
shearlet transforms, are applied and combined with deep convolutional neural network (CNN)
features. Different experiments have been conducted to investigate the number of selected
coefficients and the insertion location of the handcrafted features.
The second approach aims at reducing the segmentation algorithm’s complexity while
maintaining the segmentation performance. An attention based convolutional encode-decoder
network has been proposed to adaptively recalibrate the deep network features. This attention based
network consists of two different squeeze and excitation blocks that excite the features
spatially and channel wise. The two blocks are combined sequentially to decrease the number
of network’s parameters and reduces the model complexity. The third approach has been focuses on the application of transfer learning from different MR sequences such as T1 weighted (T1-w) and T2 weighted (T2-w) images. A
pretrained model with T1-w MR sequences is fine tuned to perform the segmentation of T2-w
images. Multiple fine tuning approaches and experiments have been conducted to study the best
fine tuning mechanism that is able to build an efficient segmentation model for both T1-w and
T2-w segmentation. Clinical datasets of fifty patients with different conditions and diagnosis have been
used to carry an objective evaluation to measure the segmentation performance of the results
obtained by the three proposed methods. The first and second approaches have been validated
with other studies in the literature that applied deep network based segmentation technique to
perform MR based attenuation correction for PET images. The proposed methods have shown
an enhancement in the bone segmentation with an increase of dice similarity coefficient (DSC)
from 0.6179 to 0.6567 using an ensemble of CNNs with an improvement percentage of 6.3%.
The proposed excitation-based CNN has decreased the model complexity by decreasing the
number of trainable parameters by more than 46% where less computing resources are required
to train the model. The proposed hybrid transfer learning method has shown its superiority to
build a multi-sequences (T1-w and T2-w) segmentation approach compared to other applied
transfer learning methods especially with the bone class where the DSC is increased from 0.3841
to 0.5393. Moreover, the hybrid transfer learning approach requires less computing time than
transfer learning using open and conservative fine tuning
Use of Coherent Point Drift in computer vision applications
This thesis presents the novel use of Coherent Point Drift in improving the robustness of a number of computer vision applications. CPD approach includes two methods for registering two images - rigid and non-rigid point set approaches which are based on the transformation model used. The key characteristic of a rigid transformation is that the distance between points is preserved, which means it can be used in the presence of translation, rotation, and scaling. Non-rigid transformations - or affine transforms - provide the opportunity of registering under non-uniform scaling and skew. The idea is to move one point set coherently to align with the second point set. The CPD method finds both the non-rigid transformation and the correspondence distance between two point sets at the same time without having to use a-priori declaration of the transformation model used.
The first part of this thesis is focused on speaker identification in video conferencing. A real-time, audio-coupled video based approach is presented, which focuses more on the video analysis side, rather than the audio analysis that is known to be prone to errors. CPD is effectively utilised for lip movement detection and a temporal face detection approach is used to minimise false positives if face detection algorithm fails to perform.
The second part of the thesis is focused on multi-exposure and multi-focus image fusion with compensation for camera shake. Scale Invariant Feature Transforms (SIFT) are first used to detect keypoints in images being fused. Subsequently this point set is reduced to remove outliers, using RANSAC (RANdom Sample Consensus) and finally the point sets are registered using CPD with non-rigid transformations. The registered images are then fused with a Contourlet based image fusion algorithm that makes use of a novel alpha blending and filtering technique to minimise artefacts. The thesis evaluates the performance of the algorithm in comparison to a number of state-of-the-art approaches, including the key commercial products available in the market at present, showing significantly improved subjective quality in the fused images.
The final part of the thesis presents a novel approach to Vehicle Make & Model Recognition in CCTV video footage. CPD is used to effectively remove skew of vehicles detected as CCTV cameras are not specifically configured for the VMMR task and may capture vehicles at different approaching angles. A LESH (Local Energy Shape Histogram) feature based approach is used for vehicle make and model recognition with the novelty that temporal processing is used to improve reliability. A number of further algorithms are used to maximise the reliability of the final outcome. Experimental results are provided to prove that the proposed system demonstrates an accuracy in excess of 95% when tested on real CCTV footage with no prior camera calibration
A Novel Multimodal Image Fusion Method Using Hybrid Wavelet-based Contourlet Transform
Various image fusion techniques have been studied to meet the requirements of different applications such as concealed weapon detection, remote sensing, urban mapping, surveillance and medical imaging. Combining two or more images of the same scene or object produces a better application-wise visible image. The conventional wavelet transform (WT) has been widely used in the field of image fusion due to its advantages, including multi-scale framework and capability of isolating discontinuities at object edges. However, the contourlet transform (CT) has been recently adopted and applied to the image fusion process to overcome the drawbacks of WT with its own advantages. Based on the experimental studies in this dissertation, it is proven that the contourlet transform is more suitable than the conventional wavelet transform in performing the image fusion. However, it is important to know that the contourlet transform also has major drawbacks. First, the contourlet transform framework does not provide shift-invariance and structural information of the source images that are necessary to enhance the fusion performance. Second, unwanted artifacts are produced during the image decomposition process via contourlet transform framework, which are caused by setting some transform coefficients to zero for nonlinear approximation. In this dissertation, a novel fusion method using hybrid wavelet-based contourlet transform (HWCT) is proposed to overcome the drawbacks of both conventional wavelet and contourlet transforms, and enhance the fusion performance. In the proposed method, Daubechies Complex Wavelet Transform (DCxWT) is employed to provide both shift-invariance and structural information, and Hybrid Directional Filter Bank (HDFB) is used to achieve less artifacts and more directional information. DCxWT provides shift-invariance which is desired during the fusion process to avoid mis-registration problem. Without the shift-invariance, source images are mis-registered and non-aligned to each other; therefore, the fusion results are significantly degraded. DCxWT also provides structural information through its imaginary part of wavelet coefficients; hence, it is possible to preserve more relevant information during the fusion process and this gives better representation of the fused image. Moreover, HDFB is applied to the fusion framework where the source images are decomposed to provide abundant directional information, less complexity, and reduced artifacts.
The proposed method is applied to five different categories of the multimodal image fusion, and experimental study is conducted to evaluate the performance of the proposed method in each multimodal fusion category using suitable quality metrics. Various datasets, fusion algorithms, pre-processing techniques and quality metrics are used for each fusion category. From every experimental study and analysis in each fusion category, the proposed method produced better fusion results than the conventional wavelet and contourlet transforms; therefore, its usefulness as a fusion method has been validated and its high performance has been verified
Automated Segmentation of Cerebral Aneurysm Using a Novel Statistical Multiresolution Approach
Cerebral Aneurysm (CA) is a vascular disease that threatens the lives of
many adults. It a ects almost 1:5 - 5% of the general population. Sub-
Arachnoid Hemorrhage (SAH), resulted by a ruptured CA, has high rates of
morbidity and mortality. Therefore, radiologists aim to detect it and diagnose
it at an early stage, by analyzing the medical images, to prevent or reduce its
damages.
The analysis process is traditionally done manually. However, with the
emerging of the technology, Computer-Aided Diagnosis (CAD) algorithms are
adopted in the clinics to overcome the traditional process disadvantages, as
the dependency of the radiologist's experience, the inter and intra observation
variability, the increase in the probability of error which increases consequently
with the growing number of medical images to be analyzed, and the artifacts
added by the medical images' acquisition methods (i.e., MRA, CTA, PET, RA,
etc.) which impedes the radiologist' s work.
Due to the aforementioned reasons, many research works propose di erent
segmentation approaches to automate the analysis process of detecting a CA
using complementary segmentation techniques; but due to the challenging task
of developing a robust reproducible reliable algorithm to detect CA regardless
of its shape, size, and location from a variety of the acquisition methods, a
diversity of proposed and developed approaches exist which still su er from
some limitations.
This thesis aims to contribute in this research area by adopting two promising
techniques based on the multiresolution and statistical approaches in the
Two-Dimensional (2D) domain. The rst technique is the Contourlet Transform
(CT), which empowers the segmentation by extracting features not apparent
in the normal image scale. While the second technique is the Hidden
Markov Random Field model with Expectation Maximization (HMRF-EM),
which segments the image based on the relationship of the neighboring pixels
in the contourlet domain.
The developed algorithm reveals promising results on the four tested Three-
Dimensional Rotational Angiography (3D RA) datasets, where an objective
and a subjective evaluation are carried out. For the objective evaluation, six
performance metrics are adopted which are: accuracy, Dice Similarity Index
(DSI), False Positive Ratio (FPR), False Negative Ratio (FNR), speci city,
and sensitivity. As for the subjective evaluation, one expert and four observers
with some medical background are involved to assess the segmentation visually.
Both evaluations compare the segmented volumes against the ground
truth data
- …