171 research outputs found
Joint Sub-component Level Segmentation and Classification for Anomaly Detection within Dual-Energy X-Ray Security Imagery
X-ray baggage security screening is in widespread use and crucial to maintaining transport security for threat/anomaly detection tasks. The automatic detection of anomaly, which is concealed within cluttered and complex electronics/electrical items, using 2D X-ray imagery is of primary interest in recent years. We address this task by introducing joint object sub-component level segmentation and classification strategy using deep Convolution Neural Network architecture. The performance is evaluated over a dataset of cluttered X-ray baggage security imagery, consisting of consumer electrical and electronics items using variants of dual-energy X-ray imagery (pseudo-colour, high, low, and effective-Z). The proposed joint sub-component level segmentation and classification approach achieve ∼ 99% true positive and ∼ 5% false positive for anomaly detection task
A Review of Automated Image Understanding within 3D Baggage Computed Tomography Security Screening
Baggage inspection is the principal safeguard against the transportation of prohibited and potentially dangerous materials at airport security checkpoints. Although traditionally performed by 2D X-ray based scanning, increasingly stringent security regulations have led to a growing demand for more advanced imaging technologies. The role of X-ray Computed Tomography is thus rapidly expanding beyond the traditional materials-based detection of explosives. The development of computer vision and image processing techniques for the automated understanding of 3D baggage-CT imagery is however, complicated by poor image resolutions, image clutter and high levels of noise and artefacts. We discuss the recent and most pertinent advancements and identify topics for future research within the challenging domain of automated image understanding for baggage security screening CT
Moving objects segmentation at a traffic junction from vehicular vision
Automatic extraction/segmentation and the recognition of moving objects on a road environment is often problematic. This is especially the case when cameras are mounted on a moving vehicle (for vehicular vision), yet this remains a critical task in vision based safety transportation. The essential problem is twofold: extracting the foreground from the moving background, and separating and recognizing pedestrians from other moving objects such as cars that appear in the foreground. The challenge of our proposed technique is to use a single mobile camera for separating the foreground from the background, and to recognize pedestrians and other objects from vehicular vision in order to achieve a low cost and intelligent driver assistance system. In this paper, the normal distribution is employed for modelling pixel gray values. The proposed technique separates the foreground from the background by comparing the pixel gray values of an input image with the normal distribution model of the pixel. The model is renewed after the separation to give a new background model for the next image. The renewal strategy changes depending on if the concerned pixel is in the background or on the foreground. Performance of the present technique was examined by real world vehicle videos captured at a junction when a car turns left or right and satisfactory results were obtained
Recommended from our members
View synthesis for kinetic depth X-ray imaging
This thesis reports the development and analysis of feature based synthesis of transmission X-ray images. The synthetic imagery is formed through matching and morphing or warping line-scan format images produced by a novel multi-view X-ray machine. In this way video type sequences, which periodically alternate between synthetic and detector based views, may be formed. The purpose of these sequences is to provide depth from motion or kinetic depth effect (KDE) in a visual display; while the role of the synthesis is to reduce the total number of detector arrays, associated collimators and X-ray flux per inspection. A specific challenge is to explore the bounds for producing synthetic imagery that can be seamlessly introduced into the resultant sequences. This work is distinct from the image collection and display technique, termed KDEX, previously undertaken by the Imaging Science Group at NTU. The ultimate aim of the research programme in collaboration with The UK Home Office and The US Dept. of Homeland Security is to enhance the detection and identification of threats in X-ray scans of luggage. A multi-view „KDEX scanner‟ was employed to collect greyscale and colour coded image sequences of 30 different bags; each sequence comprised of 7 perspective views separated from one another by 10. This imagery was organised and stored in a database to enable a coherent series of experiments to be conducted. Corresponding features in sequential pairs of images, at various different angular separations, were identified by applying a scale invariant feature transform (SIFT)
Automated Analysis of X-ray Images for Cargo Security
Customs and border officers are overwhelmed by the hundreds of millions of cargo containers that constitute the backbone of the global supply chain, any one of which could contain a security- or customs-related threat. Searching for these threats is akin to searching for needles in an ever-growing field of haystacks. This thesis considers novel automated image analysis methods to automate or assist elements of cargo inspection. The four main contributions of this thesis are as follows. Methods are proposed for the measurement and correction of detector wobble in large-scale transmission radiography using Beam Position Detectors (BPDs). Wobble is estimated from BPD measurements using a Random Regression Forest (RRF) model, Bayesian fused with a prior estimate from an Auto-Regression (AR). Next, a series of image corrections are derived, and it is shown that 87% of image error due to wobble can be corrected. This is the first proposed method for correction of wobble in large-scale transmission radiography. A Threat Image Projection (TIP) framework is proposed, for training, probing and evaluating Automated Threat Detection (ATD) algorithms. The TIP method is validated experimentally, and a method is proposed to test whether algorithms can learn to exploit TIP artefacts. A system for Empty Container Verification (ECV) is proposed. The system, trained using TIP, is based on Random Forest (RF) classification of image patches according to fixed geometric features and container location. The method outperforms previous reported results, and is able to detect very small amounts of synthetically concealed smuggled contraband. Finally, a method for ATD is proposed, based on a deep Convolutional Neural Network (CNN), trained from scratch using TIP, and exploits the material information encoded within dual-energy X-ray images to suppress false alarms. The system offers a 100-fold improvement in the false positive rate over prior work
Towards Real-Time Anomaly Detection within X-ray Security Imagery: Self-Supervised Adversarial Training Approach
Automatic threat detection is an increasingly important area in X-ray security imaging since it is critical to aid screening operators to identify concealed threats. Due to the cluttered and occluded nature of X-ray baggage imagery and limited dataset availability, few studies in the literature have systematically evaluated the automated X-ray security screening. This thesis provides an exhaustive evaluation of the use of deep Convolutional Neural Networks (CNN) for the image classification and detection problems posed within the field. The use of transfer learning overcomes the limited availability of the object of interest data examples. A thorough evaluation reveals the superiority of the CNN features over conventional hand-crafted features. Further experimentation also demonstrates the capability of the supervised deep object detection techniques as object localization strategies within cluttered X-ray security imagery. By addressing the limitations of the current X-ray datasets such as annotation and class-imbalance, the thesis subsequently transitions the scope to- wards deep unsupervised techniques for the detection of anomalies based on the training on normal (benign) X-ray samples only. The proposed anomaly detection models within the thesis employ a conditional encoder-decoder generative adversarial network that jointly learns the generation of high-dimensional image space and the inference of latent space — minimizing the distance between these images and the latent vectors during training aids in learning the data distribution for the normal samples. As a result, a larger distance metric from this learned data distribution at inference time is indicative of an outlier from that distribution — an anomaly. Experimentation over several benchmark datasets, from varying domains, shows the model efficacy and superiority over previous state-of-the-art approaches. Based on the current approaches and open problems in deep learning, the thesis finally provides discussion and future directions for X-ray security imagery
Exploiting Spatio-Temporal Coherence for Video Object Detection in Robotics
This paper proposes a method to enhance video object detection for indoor environments in robotics. Concretely, it exploits knowledge about the camera motion between frames to propagate previously detected objects to successive frames. The proposal is rooted in the concepts of planar homography to propose regions of interest where to find objects, and recursive Bayesian filtering to integrate observations over time. The proposal is evaluated on six virtual, indoor environments, accounting for the detection of nine object classes over a total of ∼ 7k frames. Results show that our proposal improves the recall and the F1-score by a factor of 1.41 and 1.27, respectively, as well as it achieves a significant reduction of the object categorization entropy (58.8%) when compared to a two-stage video object detection method used as baseline, at the cost of small time overheads (120 ms) and precision loss (0.92).</p
- …