61 research outputs found
Medical Diagnosis with Multimodal Image Fusion Techniques
Image Fusion is an effective approach utilized to draw out all the significant information from the source images, which supports experts in evaluation and quick decision making. Multi modal medical image fusion produces a composite fused image utilizing various sources to improve quality and extract complementary information. It is extremely challenging to gather every piece of information needed using just one imaging method. Therefore, images obtained from different modalities are fused Additional clinical information can be gleaned through the fusion of several types of medical image pairings. This study's main aim is to present a thorough review of medical image fusion techniques which also covers steps in fusion process, levels of fusion, various imaging modalities with their pros and cons, and the major scientific difficulties encountered in the area of medical image fusion. This paper also summarizes the quality assessments fusion metrics. The various approaches used by image fusion algorithms that are presently available in the literature are classified into four broad categories i) Spatial fusion methods ii) Multiscale Decomposition based methods iii) Neural Network based methods and iv) Fuzzy Logic based methods. the benefits and pitfalls of the existing literature are explored and Future insights are suggested. Moreover, this study is anticipated to create a solid platform for the development of better fusion techniques in medical applications
Implementation of Reversible Data Hiding Using Suitable Wavelet Transform For Controlled Contrast Enhancement
Data Hiding is important for secrete communication and it is also essential to keep the data hidden to be received by the intended recipient only. The conventional Reversible Data Hiding (RDH) algorithms pursue high Peak-Signal-to-Noise-Ratio (PSNR) at certain amount of embedding bits. Considering an importance of improvement in image visual quality than keeping high PSNR, a novel RDH scheme utilizing contrast enhancement to replace the PSNR was presented with the help of Integer Wavelet Transform (IWT). In proposed work, the identification of suitable transform from different wavelet families is planned to enhance the security of data by encrypting it and embedding more bits with the original image to generate stego image. The obtained stego image will be transmitted to the other end, where the receiver will extract the transmitted secrete data and original cover image from stego image using required keys. It will use a proper transformation for the purpose of Controlled Contrast Enhancement (CCE) to achieve the intended RDH so that the amount of embedding data bits and visual perception will be enhanced. The difference of the transmitted original image and restored original image is minor, which is almost invisible for human eyes though more bits are embedded with the original image. The performance parameters are also calculated
Automatic Side-Scan Sonar Image Enhancement in Curvelet Transform Domain
We propose a novel automatic side-scan sonar image enhancement algorithm based on curvelet transform. The proposed algorithm uses the curvelet transform to construct a multichannel enhancement structure based on human visual system (HVS) and adopts a new adaptive nonlinear mapping scheme to modify the curvelet transform coefficients in each channel independently and automatically. Firstly, the noisy and low-contrast sonar image is decomposed into a low frequency channel and a series of high frequency channels by using curvelet transform. Secondly, a new nonlinear mapping scheme, which coincides with the logarithmic nonlinear enhancement characteristic of the HVS perception, is designed without any parameter tuning to adjust the curvelet transform coefficients in each channel. Finally, the enhanced image can be reconstructed with the modified coefficients via inverse curvelet transform. The enhancement is achieved by amplifying subtle features, improving contrast, and eliminating noise simultaneously. Experiment results show that the proposed algorithm produces better enhanced results than state-of-the-art algorithms
Survey on wavelet based image fusion techniques
Image fusion is the process of combining multiple images into a single image without distortion or loss of information. The techniques related to image fusion are broadly classified as spatial and transform domain methods. In which, the transform domain based wavelet fusion techniques are widely used in different domains like medical, space and military for the fusion of multimodality or multi-focus images. In this paper, an overview of different wavelet transform based methods and its applications for image fusion are discussed and analysed
Recommended from our members
Deep learning assisted MRI guided attenuation correction in PET
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University LondonPositron emission tomography (PET) is a unique imaging modality that provides physiological
and functional details of the tissue at the molecular level. However, the acquired PET images
have some limitations such as the attenuation. PET attenuation correction is an essential step to
obtain the full potential of PET quantification. With the wide use of hybrid PET/MR scanners,
magnetic resonance (MR) images are used to address the problem of PET attenuation correction.
The MR images segmentation is one simple and robust approach to create pseudo computed
tomography (CT) images, which are used to generate attenuation coefficient maps to correct the
PET attenuation. Recently, deep learning has been proposed and used as a promising technique
to efficiently perform MR and various medical images segmentation.
In this research work, deep learning guided segmentation approaches have been proposed
to enhance the bone class segmentation of MR brain images in order to generate accurate
pseudo-CT images. The first approach has introduced the combination of handcrafted features
with deep learning features to enrich the set of features. Multiresolution analysis techniques,
which generate multiscale and multidirectional coefficients of an image such as contourlet and
shearlet transforms, are applied and combined with deep convolutional neural network (CNN)
features. Different experiments have been conducted to investigate the number of selected
coefficients and the insertion location of the handcrafted features.
The second approach aims at reducing the segmentation algorithm’s complexity while
maintaining the segmentation performance. An attention based convolutional encode-decoder
network has been proposed to adaptively recalibrate the deep network features. This attention based
network consists of two different squeeze and excitation blocks that excite the features
spatially and channel wise. The two blocks are combined sequentially to decrease the number
of network’s parameters and reduces the model complexity. The third approach has been focuses on the application of transfer learning from different MR sequences such as T1 weighted (T1-w) and T2 weighted (T2-w) images. A
pretrained model with T1-w MR sequences is fine tuned to perform the segmentation of T2-w
images. Multiple fine tuning approaches and experiments have been conducted to study the best
fine tuning mechanism that is able to build an efficient segmentation model for both T1-w and
T2-w segmentation. Clinical datasets of fifty patients with different conditions and diagnosis have been
used to carry an objective evaluation to measure the segmentation performance of the results
obtained by the three proposed methods. The first and second approaches have been validated
with other studies in the literature that applied deep network based segmentation technique to
perform MR based attenuation correction for PET images. The proposed methods have shown
an enhancement in the bone segmentation with an increase of dice similarity coefficient (DSC)
from 0.6179 to 0.6567 using an ensemble of CNNs with an improvement percentage of 6.3%.
The proposed excitation-based CNN has decreased the model complexity by decreasing the
number of trainable parameters by more than 46% where less computing resources are required
to train the model. The proposed hybrid transfer learning method has shown its superiority to
build a multi-sequences (T1-w and T2-w) segmentation approach compared to other applied
transfer learning methods especially with the bone class where the DSC is increased from 0.3841
to 0.5393. Moreover, the hybrid transfer learning approach requires less computing time than
transfer learning using open and conservative fine tuning
Bidirectional ConvLSTMXNet for Brain Tumor Segmentation of MR Images
In recent years, deep learning based networks have achieved good performance in brain tumour segmentation of MR Image. Among the existing networks, U-Net has been successfully applied. In this paper, it is propose deep-learning based Bidirectional Convolutional LSTM XNet (BConvLSTMXNet) for segmentation of brain tumor and using GoogLeNet classify tumor & non-tumor. Evaluated on BRATS-2019 data-set and the results are obtained for classification of tumor and non-tumor with Accuracy: 0.91, Precision: 0.95, Recall: 1.00 & F1-Score: 0.92. Similarly for segmentation of brain tumor obtained Accuracy: 0.99, Specificity: 0.98, Sensitivity: 0.91, Precision: 0.91 & F1-Score: 0.88
- …