31 research outputs found
Multispectral Palmprint Encoding and Recognition
Palmprints are emerging as a new entity in multi-modal biometrics for human
identification and verification. Multispectral palmprint images captured in the
visible and infrared spectrum not only contain the wrinkles and ridge structure
of a palm, but also the underlying pattern of veins; making them a highly
discriminating biometric identifier. In this paper, we propose a feature
encoding scheme for robust and highly accurate representation and matching of
multispectral palmprints. To facilitate compact storage of the feature, we
design a binary hash table structure that allows for efficient matching in
large databases. Comprehensive experiments for both identification and
verification scenarios are performed on two public datasets -- one captured
with a contact-based sensor (PolyU dataset), and the other with a contact-free
sensor (CASIA dataset). Recognition results in various experimental setups show
that the proposed method consistently outperforms existing state-of-the-art
methods. Error rates achieved by our method (0.003% on PolyU and 0.2% on CASIA)
are the lowest reported in literature on both dataset and clearly indicate the
viability of palmprint as a reliable and promising biometric. All source codes
are publicly available.Comment: Preliminary version of this manuscript was published in ICCV 2011. Z.
Khan A. Mian and Y. Hu, "Contour Code: Robust and Efficient Multispectral
Palmprint Encoding for Human Recognition", International Conference on
Computer Vision, 2011. MATLAB Code available:
https://sites.google.com/site/zohaibnet/Home/code
Multisensor image fusion approach utilizing hybrid pre-enhancement and double nonsubsampled contourlet transform
A multisensor image fusion approach established on the hybrid-domain image enhancement and double nonsubsampled contourlet transform (NSCT) is proposed. The hybrid-domain pre-enhancement algorithm can promote the contrast of the visible color image. Different fusion rules are, respectively, selected and applied to obtain fusion results. The double NSCT framework is introduced to obtain better fusion performance than the general single NSCT framework. Experimental outcomes in fused images and performance results demonstrate that the presented approach is apparently more advantageous
Breast Mass Segmentation Using a Semi-automatic Procedure Based on Fuzzy C-means Clustering
Mammography is the primary modality that helped in the early detection and diagnosis of women breast diseases. Further, the process of extracting the masses in mammogram represents a challenging task facing the radiologists, due to problems such as fuzzy or speculated borders, low contrast and the presence of intensity inhomogeneities. Aims to help the radiologists in the diagnosis of breast cancer, many approaches have been conducted to automatically segment the masses in mammograms. Towards this aim, in this paper, we present a new approach for extraction of tumors from region-of-interest (ROI) using the algorithm of Fuzzy C-Means (FCM) setting two clusters for semi-automated segmentation. The proposed method meant to select as input data the set of pixels that enable to get the meaningful information required to segment the masses with high accuracy. This could be accomplished through eliminating unnecessary pixels, which influence on this process through separating it outside of the input data using an optimal threshold given by monitoring the change of clusters rate during the process of threshold decrementing. The proposed methodology has successfully segmented the masses, with an average sensitivity of 82.02% and specificity of 98.23%
A novel multispectral and 2.5D/3D image fusion camera system for enhanced face recognition
The fusion of images from the visible and long-wave infrared (thermal) portions of the spectrum
produces images that have improved face recognition performance under varying lighting conditions.
This is because long-wave infrared images are the result of emitted, rather than reflected,
light and are therefore less sensitive to changes in ambient light. Similarly, 3D and 2.5D images
have also improved face recognition under varying pose and lighting. The opacity of glass to
long-wave infrared light, however, means that the presence of eyeglasses in a face image reduces
the recognition performance.
This thesis presents the design and performance evaluation of a novel camera system which is
capable of capturing spatially registered visible, near-infrared, long-wave infrared and 2.5D depth
video images via a common optical path requiring no spatial registration between sensors beyond
scaling for differences in sensor sizes. Experiments using a range of established face recognition
methods and multi-class SVM classifiers show that the fused output from our camera system not
only outperforms the single modality images for face recognition, but that the adaptive fusion
methods used produce consistent increases in recognition accuracy under varying pose, lighting
and with the presence of eyeglasses
A New Approach to the Detection of Mammogram Boundary
Mammography is a method used for the detection of breast cancer. computer-aided diagnostic (CAD) systems help the radiologist in the detection and interpretation of mass in breast mammography. One of the important information of a mass is its contour and its form because it provides valuable information about the abnormality of a mass. The accuracy in the recognition of the shape of a mass is related to the accuracy of the detected mass contours. In this work we propose a new approach for detecting the boundaries of lesion in mammography images based on region growing algorithm without using the threshold, the proposed method requires an initial rectangle surrounding the lesion selected manually by the radiologist (Region Of Interest), where the region growing algorithm applies on lines segments that attach each pixel of this rectangle with the seed point, such as the ends (seeds) of each line segment grow in a direction towards one another. The proposed approach is evaluated on a set of data with 20 masses of the MIAS base whose contours are annotated manually by expert radiologists. The performance of the method is evaluated in terms of specificity, sensitivity, accuracy and overlap. All the findings and details of approach are presented in detail
A novel multispectral and 2.5D/3D image fusion camera system for enhanced face recognition
The fusion of images from the visible and long-wave infrared (thermal) portions of the spectrum
produces images that have improved face recognition performance under varying lighting conditions.
This is because long-wave infrared images are the result of emitted, rather than reflected,
light and are therefore less sensitive to changes in ambient light. Similarly, 3D and 2.5D images
have also improved face recognition under varying pose and lighting. The opacity of glass to
long-wave infrared light, however, means that the presence of eyeglasses in a face image reduces
the recognition performance.
This thesis presents the design and performance evaluation of a novel camera system which is
capable of capturing spatially registered visible, near-infrared, long-wave infrared and 2.5D depth
video images via a common optical path requiring no spatial registration between sensors beyond
scaling for differences in sensor sizes. Experiments using a range of established face recognition
methods and multi-class SVM classifiers show that the fused output from our camera system not
only outperforms the single modality images for face recognition, but that the adaptive fusion
methods used produce consistent increases in recognition accuracy under varying pose, lighting
and with the presence of eyeglasses
Accurate Iris Localization Using Edge Map Generation and Adaptive Circular Hough Transform for Less Constrained Iris Images
This paper proposes an accurate iris localization algorithm for the iris images acquired under near infrared (NIR) illuminations and having noise due to eyelids, eyelashes, lighting reflections, non-uniform illumination, eyeglasses and eyebrow hair etc. The two main contributions in the paper are an edge map generation technique for pupil boundary detection and an adaptive circular Hough transform (CHT) algorithm for limbic boundary detection, which not only make the iris localization more accurate but faster also. The edge map for pupil boundary detection is generated on intersection (logical AND) of two binary edge maps obtained using thresholding, morphological operations and Sobel edge detection, which results in minimal false edges caused by the noise. The adaptive CHT algorithm for limbic boundary detection searches for a set of two arcs in an image instead of a full circle that counters iris-occlusions by the eyelids and eyelashes. The proposed CHT and adaptive CHT implementations for pupil and limbic boundary detection respectively use a two-dimensional accumulator array that reduces memory requirements. The proposed algorithm gives the accuracies of 99.7% and 99.38% for the challenging CASIA-Iris-Thousand (version 4.0) and CASIA-Iris-Lamp (version 3.0) databases respectively. The average time cost per image is 905 msec. The proposed algorithm is compared with the previous work and shows better results
Finger Vein Image Enhancement Technique based on Gabor filter and Discrete Cosine Transform
Biometrics is a global technique to establish the identity of a person by measuring one of their physical or behavioral characteristics such as fingerprint, signature, iris, voice and face. Compared to these biometric techniques, the finger vein technique has distinct advantages as it helps to protect privacy and anonymity in automated individual recognition. Many studies showed that the finger vein images were of a low quality because of the variation in the tissues and uneven illumination. Hence, there is a need for effective image enhancement techniques, which can improve the quality of the images. In this study, we proposed a novel technique, which enhances the image quality of the finger veins. This method includes contrast amelioration, use of Gabor filters and image fusion, which generates an image with highly connective patterns. We used three criteria to evaluate the quality of processed images, the mean of grey values, the image entropy, and the image contrast. The obtained result shows higher values when using our approach in comparison to the baseline methods considered in this work
Recommended from our members
Deep learning assisted MRI guided attenuation correction in PET
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University LondonPositron emission tomography (PET) is a unique imaging modality that provides physiological
and functional details of the tissue at the molecular level. However, the acquired PET images
have some limitations such as the attenuation. PET attenuation correction is an essential step to
obtain the full potential of PET quantification. With the wide use of hybrid PET/MR scanners,
magnetic resonance (MR) images are used to address the problem of PET attenuation correction.
The MR images segmentation is one simple and robust approach to create pseudo computed
tomography (CT) images, which are used to generate attenuation coefficient maps to correct the
PET attenuation. Recently, deep learning has been proposed and used as a promising technique
to efficiently perform MR and various medical images segmentation.
In this research work, deep learning guided segmentation approaches have been proposed
to enhance the bone class segmentation of MR brain images in order to generate accurate
pseudo-CT images. The first approach has introduced the combination of handcrafted features
with deep learning features to enrich the set of features. Multiresolution analysis techniques,
which generate multiscale and multidirectional coefficients of an image such as contourlet and
shearlet transforms, are applied and combined with deep convolutional neural network (CNN)
features. Different experiments have been conducted to investigate the number of selected
coefficients and the insertion location of the handcrafted features.
The second approach aims at reducing the segmentation algorithm’s complexity while
maintaining the segmentation performance. An attention based convolutional encode-decoder
network has been proposed to adaptively recalibrate the deep network features. This attention based
network consists of two different squeeze and excitation blocks that excite the features
spatially and channel wise. The two blocks are combined sequentially to decrease the number
of network’s parameters and reduces the model complexity. The third approach has been focuses on the application of transfer learning from different MR sequences such as T1 weighted (T1-w) and T2 weighted (T2-w) images. A
pretrained model with T1-w MR sequences is fine tuned to perform the segmentation of T2-w
images. Multiple fine tuning approaches and experiments have been conducted to study the best
fine tuning mechanism that is able to build an efficient segmentation model for both T1-w and
T2-w segmentation. Clinical datasets of fifty patients with different conditions and diagnosis have been
used to carry an objective evaluation to measure the segmentation performance of the results
obtained by the three proposed methods. The first and second approaches have been validated
with other studies in the literature that applied deep network based segmentation technique to
perform MR based attenuation correction for PET images. The proposed methods have shown
an enhancement in the bone segmentation with an increase of dice similarity coefficient (DSC)
from 0.6179 to 0.6567 using an ensemble of CNNs with an improvement percentage of 6.3%.
The proposed excitation-based CNN has decreased the model complexity by decreasing the
number of trainable parameters by more than 46% where less computing resources are required
to train the model. The proposed hybrid transfer learning method has shown its superiority to
build a multi-sequences (T1-w and T2-w) segmentation approach compared to other applied
transfer learning methods especially with the bone class where the DSC is increased from 0.3841
to 0.5393. Moreover, the hybrid transfer learning approach requires less computing time than
transfer learning using open and conservative fine tuning