191 research outputs found
Hand-Based Biometric Analysis
Hand-based biometric analysis systems and techniques provide robust hand-based identification and verification. An image of a hand is obtained, which is then segmented into a palm region and separate finger regions. Acquisition of the image is performed without requiring particular orientation or placement restrictions. Segmentation is performed without the use of reference points on the images. Each segment is analyzed by calculating a set of Zernike moment descriptors for the segment. The feature parameters thus obtained are then fused and compared to stored sets of descriptors in enrollment templates to arrive at an identity decision. By using Zernike moments, and through additional manipulation, the biometric analysis is invariant to rotation, scale, or translation or an input image. Additionally, the analysis uses re-use of commonly seen terms in Zernike calculations to achieve additional efficiencies over traditional Zernike moment calculation
Hand-Based Biometric Analysis
Hand-based biometric analysis systems and techniques are described which provide robust hand-based identification and verification. An image of a hand is obtained, which is then segmented into a palm region and separate finger regions. Acquisition of the image is performed without requiring particular orientation or placement restrictions. Segmentation is performed without the use of reference points on the images. Each segment is analyzed by calculating a set of Zernike moment descriptors for the segment. The feature parameters thus obtained are then fused and compared to stored sets of descriptors in enrollment templates to arrive at an identity decision. By using Zernike moments, and through additional manipulation, the biometric analysis is invariant to rotation, scale, or translation or an in put image. Additionally, the analysis utilizes re-use of commonly-seen terms in Zernike calculations to achieve additional efficiencies over traditional Zernike moment calculation
ConnectedUNets++: Mass Segmentation from Whole Mammographic Images
Deep learning has made a breakthrough in medical image segmentation in recent
years due to its ability to extract high-level features without the need for
prior knowledge. In this context, U-Net is one of the most advanced medical
image segmentation models, with promising results in mammography. Despite its
excellent overall performance in segmenting multimodal medical images, the
traditional U-Net structure appears to be inadequate in various ways. There are
certain U-Net design modifications, such as MultiResUNet, Connected-UNets, and
AU-Net, that have improved overall performance in areas where the conventional
U-Net architecture appears to be deficient. Following the success of UNet and
its variants, we have presented two enhanced versions of the Connected-UNets
architecture: ConnectedUNets+ and ConnectedUNets++. In ConnectedUNets+, we have
replaced the simple skip connections of Connected-UNets architecture with
residual skip connections, while in ConnectedUNets++, we have modified the
encoder-decoder structure along with employing residual skip connections. We
have evaluated our proposed architectures on two publicly available datasets,
the Curated Breast Imaging Subset of Digital Database for Screening Mammography
(CBIS-DDSM) and INbreast.Comment: Results are to be update
Comparison of Semantic Segmentation Approaches for Horizon/Sky Line Detection
Horizon or skyline detection plays a vital role towards mountainous visual
geo-localization, however most of the recently proposed visual geo-localization
approaches rely on \textbf{user-in-the-loop} skyline detection methods.
Detecting such a segmenting boundary fully autonomously would definitely be a
step forward for these localization approaches. This paper provides a
quantitative comparison of four such methods for autonomous horizon/sky line
detection on an extensive data set. Specifically, we provide the comparison
between four recently proposed segmentation methods; one explicitly targeting
the problem of horizon detection\cite{Ahmad15}, second focused on visual
geo-localization but relying on accurate detection of skyline \cite{Saurer16}
and other two proposed for general semantic segmentation -- Fully Convolutional
Networks (FCN) \cite{Long15} and SegNet\cite{Badrinarayanan15}. Each of the
first two methods is trained on a common training set \cite{Baatz12} comprised
of about 200 images while models for the third and fourth method are fine tuned
for sky segmentation problem through transfer learning using the same data set.
Each of the method is tested on an extensive test set (about 3K images)
covering various challenging geographical, weather, illumination and seasonal
conditions. We report average accuracy and average absolute pixel error for
each of the presented formulation.Comment: Proceedings of the International Joint Conference on Neural Networks
(IJCNN) (oral presentation), IEEE Computational Intelligence Society, 201
- …