2 research outputs found
Deep Learning-Based Approaches for Contactless Fingerprints Segmentation and Extraction
Fingerprints are widely recognized as one of the most unique and reliable
characteristics of human identity. Most modern fingerprint authentication
systems rely on contact-based fingerprints, which require the use of
fingerprint scanners or fingerprint sensors for capturing fingerprints during
the authentication process. Various types of fingerprint sensors, such as
optical, capacitive, and ultrasonic sensors, employ distinct techniques to
gather and analyze fingerprint data. This dependency on specific hardware or
sensors creates a barrier or challenge for the broader adoption of fingerprint
based biometric systems. This limitation hinders the widespread adoption of
fingerprint authentication in various applications and scenarios. Border
control, healthcare systems, educational institutions, financial transactions,
and airport security face challenges when fingerprint sensors are not
universally available. To mitigate the dependence on additional hardware, the
use of contactless fingerprints has emerged as an alternative. Developing
precise fingerprint segmentation methods, accurate fingerprint extraction
tools, and reliable fingerprint matchers are crucial for the successful
implementation of a robust contactless fingerprint authentication system. This
paper focuses on the development of a deep learning-based segmentation tool for
contactless fingerprint localization and segmentation. Our system leverages
deep learning techniques to achieve high segmentation accuracy and reliable
extraction of fingerprints from contactless fingerprint images. In our
evaluation, our segmentation method demonstrated an average mean absolute error
(MAE) of 30 pixels, an error in angle prediction (EAP) of 5.92 degrees, and a
labeling accuracy of 97.46%. These results demonstrate the effectiveness of our
novel contactless fingerprint segmentation and extraction tools
Vision based intelligent traffic light management system using Faster R-CNN
Transportation systems primarily depend on vehicular flow on roads. Developed countries have shifted towards automated signal control, which manages and updates signal synchronisation automatically. In contrast, traffic in underdeveloped countries is mainly governed by manual traffic light systems. These existing manual systems lead to numerous issues, wasting substantial resources such as time, energy, and fuel, as they cannot make real-time decisions. In this work, we propose an algorithm to determine traffic signal durations based on real-time vehicle density, obtained from live closed circuit television camera feeds adjacent to traffic signals. The algorithm automates the traffic light system, making decisions based on vehicle density and employing Faster R-CNN for vehicle detection. Additionally, we have created a local dataset from live streams of Punjab Safe City cameras in collaboration with the local police authority. The proposed algorithm achieves a class accuracy of 96.6% and a vehicle detection accuracy of 95.7%. Across both day and night modes, our proposed method maintains an average precision, recall, F1 score, and vehicle detection accuracy of 0.94, 0.98, 0.96 and 0.95, respectively. Our proposed work surpasses all evaluation metrics compared to state-of-the-art methodologies