978 research outputs found
Automatic human behaviour anomaly detection in surveillance video
This thesis work focusses upon developing the capability to automatically evaluate
and detect anomalies in human behaviour from surveillance video. We work with
static monocular cameras in crowded urban surveillance scenarios, particularly air-
ports and commercial shopping areas. Typically a person is 100 to 200 pixels high
in a scene ranging from 10 - 20 meters width and depth, populated by 5 to 40 peo-
ple at any given time. Our procedure evaluates human behaviour unobtrusively to
determine outlying behavioural events,
agging abnormal events to the operator.
In order to achieve automatic human behaviour anomaly detection we address
the challenge of interpreting behaviour within the context of the social and physical
environment. We develop and evaluate a process for measuring social connectivity
between individuals in a scene using motion and visual attention features. To do this
we use mutual information and Euclidean distance to build a social similarity matrix
which encodes the social connection strength between any two individuals. We de-
velop a second contextual basis which acts by segmenting a surveillance environment
into behaviourally homogeneous subregions which represent high tra c slow regions
and queuing areas. We model the heterogeneous scene in homogeneous subgroups
using both contextual elements. We bring the social contextual information, the
scene context, the motion, and visual attention features together to demonstrate
a novel human behaviour anomaly detection process which nds outlier behaviour
from a short sequence of video. The method, Nearest Neighbour Ranked Outlier
Clusters (NN-RCO), is based upon modelling behaviour as a time independent se-
quence of behaviour events, can be trained in advance or set upon a single sequence.
We nd that in a crowded scene the application of Mutual Information-based social
context permits the ability to prevent self-justifying groups and propagate anomalies
in a social network, granting a greater anomaly detection capability. Scene context
uniformly improves the detection of anomalies in all the datasets we test upon.
We additionally demonstrate that our work is applicable to other data domains.
We demonstrate upon the Automatic Identi cation Signal data in the maritime
domain. Our work is capable of identifying abnormal shipping behaviour using joint
motion dependency as analogous for social connectivity, and similarly segmenting
the shipping environment into homogeneous regions
Automatic Pipeline Surveillance Air-Vehicle
This thesis presents the developments of a vision-based system for
aerial pipeline Right-of-Way surveillance using optical/Infrared sensors mounted
on Unmanned Aerial Vehicles (UAV). The aim of research is to develop a highly
automated, on-board system for detecting and following the pipelines; while
simultaneously detecting any third-party interference. The proposed approach
of using a UAV platform could potentially reduce the cost of monitoring and
surveying pipelines when compared to manned aircraft. The main contributions
of this thesis are the development of the image-analysis algorithms, the overall
system architecture and validation of in hardware based on scaled down Test
environment.
To evaluate the performance of the system, the algorithms were coded using
Python programming language. A small-scale test-rig of the pipeline structure,
as well as expected third-party interference, was setup to simulate the
operational environment and capture/record data for the algorithm testing and
validation.
The pipeline endpoints are identified by transforming the 16-bits depth data of
the explored environment into 3D point clouds world coordinates. Then, using
the Random Sample Consensus (RANSAC) approach, the foreground and
background are separated based on the transformed 3D point cloud to extract
the plane that corresponds to the ground. Simultaneously, the boundaries of the
explored environment are detected based on the 16-bit depth data using a
canny detector. Following that, these boundaries were filtered out, after being
transformed into a 3D point cloud, based on the real height of the pipeline for fast and accurate measurements using a Euclidean distance of each boundary
point, relative to the plane of the ground extracted previously. The filtered
boundaries were used to detect the straight lines of the object boundary (Hough
lines), once transformed into 16-bit depth data, using a Hough transform
method. The pipeline is verified by estimating a centre line segment, using a 3D
point cloud of each pair of the Hough line segments, (transformed into 3D).
Then, the corresponding linearity of the pipeline points cloud is filtered within
the width of the pipeline using Euclidean distance in the foreground point cloud.
Then, the segment length of the detected centre line is enhanced to match the
exact pipeline segment by extending it along the filtered point cloud of the
pipeline.
The third-party interference is detected based on four parameters, namely:
foreground depth data; pipeline depth data; pipeline endpoints location in the
3D point cloud; and Right-of-Way distance. The techniques include detection,
classification, and localization algorithms.
Finally, a waypoints-based navigation system was implemented for the air-
vehicle to fly over the course waypoints that were generated online by a
heading angle demand to follow the pipeline structure in real-time based on the
online identification of the pipeline endpoints relative to a camera frame
Toward Sensor Modular Autonomy for Persistent Land Intelligence Surveillance and Reconnaissance (ISR)
Currently, most land Intelligence, Surveillance and Reconnaissance (ISR) assets (e.g. EO/IR cameras) are simply data collectors. Understanding, decision making and sensor control are performed by the human operators, involving high cognitive load. Any automation in the system has traditionally involved bespoke design of centralised systems that are highly specific for the assets/targets/environment under consideration, resulting in complex, non-flexible systems that exhibit poor interoperability. We address a concept of Autonomous Sensor Modules (ASMs) for land ISR, where these modules have the ability to make low-level decisions on their own in order to fulfil a higher-level objective, and plug in, with the minimum of preconfiguration, to a High Level Decision Making Module (HLDMM) through a middleware integration layer. The dual requisites of autonomy and interoperability create challenges around information fusion and asset management in an autonomous hierarchical system, which are addressed in this work. This paper presents the results of a demonstration system, known as Sensing for Asset Protection with Integrated Electronic Networked Technology (SAPIENT), which was shown in realistic base protection scenarios with live sensors and targets. The SAPIENT system performed sensor cueing, intelligent fusion, sensor tasking, target hand-off and compensation for compromised sensors, without human control, and enabled rapid integration of ISR assets at the time of system deployment, rather than at design-time. Potential benefits include rapid interoperability for coalition operations, situation understanding with low operator cognitive burden and autonomous sensor management in heterogenous sensor systems
Autonomous computational intelligence-based behaviour recognition in security and surveillance
This paper presents a novel approach to sensing both suspicious, and task-specific behaviours through the use of advanced computational intelligence techniques. Locating suspicious activity in surveillance camera networks is an intensive task due to the volume of information and large numbers of camera sources to monitor. This results in countless hours of video data being streamed to disk without being screened by a human operator. To address this need, there are emerging video analytics solutions that have introduced new metrics such as people counting and route monitoring, alongside more traditional alerts such as motion detection. There are however few solutions that are sufficiently robust to reduce the need for human operators in these environments, and new approaches are needed to address the uncertainty in identifying and classifying human behaviours, autonomously, from a video stream. In this work we present an approach to address the autonomous identification of human behaviours derived from human pose analysis. Behavioural recognition is a significant challenge due to the complex subtleties that often make up an action; the large overlap in cues results in high levels of classification uncertainty. False alarms are significant impairments to autonomous detection and alerting systems, and over reporting can lead to systems being muted, disabled, or decommissioned. We present results on a Computational-Intelligence based Behaviour Recognition (CIBR) that utilises artificial intelligence to learn, optimise, and classify human activity. We achieve this through extraction of skeleton recognition of human forms within an image. A type-2 Fuzzy logic classifier then converts the human skeletal forms into a set of base atomic poses (standing, walking, etc.), after which a Markov-chain model is used to order a pose sequence. Through this method we are able to identify, with good accuracy, several classes of human behaviour that correlate with known suspicious, or anomalous, behaviours
Robust abandoned object detection integrating wide area visual surveillance and social context
This paper presents a video surveillance framework that robustly and efficiently detects abandoned objects in surveillance scenes. The framework is based on a novel threat assessment algorithm which combines the concept of ownership with automatic understanding of social relations in order to infer abandonment of objects. Implementation is achieved through development of a logic-based inference engine based on Prolog. Threat detection performance is conducted by testing against a range of datasets describing realistic situations and demonstrates a reduction in the number of false alarms generated. The proposed system represents the approach employed in the EU SUBITO project (Surveillance of Unattended Baggage and the Identification and Tracking of the Owner)
See no Evil: Challenges of security surveillance and monitoring
While the development of intelligent technologies in security surveillance can augment human capabilities, they do not replace the role of the operator entirely; as such, when developing surveillance support it is critical that limitations to the cognitive system are taken into account. The current article reviews the cognitive challenges associated with the task of a CCTV operator: visual search and cognitive/perceptual overload, attentional failures, vulnerability to distraction, and decision-making in a dynamically evolving environment. While not directly applied to surveillance issues, we suggest that the NSEEV (noticing – salience, effort, expectancy, value) model of attention could provide a useful theoretical basis for understanding the challenges faced in detection and monitoring tasks. Having identified cognitive limitations of the human operator, this review sets out a research agenda for further understanding the cognitive functioning related to surveillance, and highlights the need to consider the human element at the design stage when developing technological solutions to security surveillance
- …