36 research outputs found
Novelty detection in video surveillance using hierarchical neural networks
Abstract. A hierarchical self-organising neural network is described for the detection of unusual pedestrian behaviour in video-based surveillance systems. The system is trained on a normal data set, with no prior information about the
scene under surveillance, thereby requiring minimal user input. Nodes use a trace activation rule and feedforward connections, modified so that higher layer nodes are sensitive to trajectory segments traced across the previous layer. Top layer nodes have binary lateral connections and corresponding ânovelty accumulatorâ nodes. Lateral connections are set between co-occurring nodes, generating a signal to prevent accumulation of the novelty measure along normal sequences. In abnormal sequences the novelty accumulator nodes are allowed to increase their activity, generating an alarm state
A Neural System for Automated CCTV Surveillance
This paper overviews a new system, the âOwens
Tracker,â for automated identification of suspicious
pedestrian activity in a car-park.
Centralized CCTV systems relay multiple video streams
to a central point for monitoring by an operator. The
operator receives a continuous stream of information,
mostly related to normal activity, making it difficult to
maintain concentration at a sufficiently high level.
While it is difficult to place quantitative boundaries on
the number of scenes and time period over which
effective monitoring can be performed, Wallace and
Diffley [1] give some guidance, based on empirical and
anecdotal evidence, suggesting that the number of
cameras monitored by an operator be no greater than 16,
and that the period of effective monitoring may be as
low as 30 minutes before recuperation is required.
An intelligent video surveillance system should
therefore act as a filter, censuring inactive scenes and
scenes showing normal activity. By presenting the
operator only with unusual activity his/her attention is
effectively focussed, and the ratio of cameras to
operators can be increased.
The Owens Tracker learns to recognize environmentspecific
normal behaviour, and refers sequences of
unusual behaviour for operator attention. The system
was developed using standard low-resolution CCTV
cameras operating in the car-parks of Doxford Park
Industrial Estate (Sunderland, Tyne and Wear), and
targets unusual pedestrian behaviour.
The modus operandi of the system is to highlight
excursions from a learned model of normal behaviour in
the monitored scene. The system tracks objects and
extracts their centroids; behaviour is defined as the
trajectory traced by an object centroid; normality as the
trajectories typically encountered in the scene. The
essential stages in the system are: segmentation of
objects of interest; disambiguation and tracking of
multiple contacts, including the handling of occlusion
and noise, and successful tracking of objects that
âmergeâ during motion; identification of unusual
trajectories. These three stages are discussed in more
detail in the following sections, and the system
performance is then evaluated
A neural system for automated CCTV surveillance
This paper overviews a new system, the âOwens Tracker, â for automated identification of suspicious pedestrian activity in a car-park. Centralized CCTV systems relay multiple video stream
Assessing similarity of dynamic geographic phenomena in spatiotemporal databases.
The growing availability of routine observations from satellite imagery and other remote sensors holds great promise for improved understanding of processes that act in the landscape. However, geographers' ability to effectively use such spatiotemporal data is challenged by large data volume and limitations of conventional data models in geographic information systems (GIS), which provide limited support for querying and exploration of spatiotemporal data other than simple comparisons of temporally referenced snapshots. Current GIS representations allow measurement of change but do not address coherent patterns of change that reflects the working of geographic events and processes. This dissertation presents a representational and query framework to overcome the limitations and enable assessing similarity of dynamic phenomena. The research includes three self contained but related studies: (1) development of a representational framework that incorporates spatiotemporal properties of geographic phenomena, (2) development of a framework to characterize events and processes that can be inferred from GIS databases, and (3) development of a method to assess similarity of events and processes based on the temporal sequences of spatiotemporal properties. Collectively the studies contribute to scientific understanding of spatiotemporal components of geographic processes and technological advances in representation and analysis
Urban Informatics
This open access book is the first to systematically introduce the principles of urban informatics and its application to every aspect of the city that involves its functioning, control, management, and future planning. It introduces new models and tools being developed to understand and implement these technologies that enable cities to function more efficiently â to become âsmartâ and âsustainableâ. The smart city has quickly emerged as computers have become ever smaller to the point where they can be embedded into the very fabric of the city, as well as being central to new ways in which the population can communicate and act. When cities are wired in this way, they have the potential to become sentient and responsive, generating massive streams of âbigâ data in real time as well as providing immense opportunities for extracting new forms of urban data through crowdsourcing. This book offers a comprehensive review of the methods that form the core of urban informatics from various kinds of urban remote sensing to new approaches to machine learning and statistical modelling. It provides a detailed technical introduction to the wide array of tools information scientists need to develop the key urban analytics that are fundamental to learning about the smart city, and it outlines ways in which these tools can be used to inform design and policy so that cities can become more efficient with a greater concern for environment and equity
Urban Informatics
This open access book is the first to systematically introduce the principles of urban informatics and its application to every aspect of the city that involves its functioning, control, management, and future planning. It introduces new models and tools being developed to understand and implement these technologies that enable cities to function more efficiently â to become âsmartâ and âsustainableâ. The smart city has quickly emerged as computers have become ever smaller to the point where they can be embedded into the very fabric of the city, as well as being central to new ways in which the population can communicate and act. When cities are wired in this way, they have the potential to become sentient and responsive, generating massive streams of âbigâ data in real time as well as providing immense opportunities for extracting new forms of urban data through crowdsourcing. This book offers a comprehensive review of the methods that form the core of urban informatics from various kinds of urban remote sensing to new approaches to machine learning and statistical modelling. It provides a detailed technical introduction to the wide array of tools information scientists need to develop the key urban analytics that are fundamental to learning about the smart city, and it outlines ways in which these tools can be used to inform design and policy so that cities can become more efficient with a greater concern for environment and equity
Urban Informatics
This open access book is the first to systematically introduce the principles of urban informatics and its application to every aspect of the city that involves its functioning, control, management, and future planning. It introduces new models and tools being developed to understand and implement these technologies that enable cities to function more efficiently â to become âsmartâ and âsustainableâ. The smart city has quickly emerged as computers have become ever smaller to the point where they can be embedded into the very fabric of the city, as well as being central to new ways in which the population can communicate and act. When cities are wired in this way, they have the potential to become sentient and responsive, generating massive streams of âbigâ data in real time as well as providing immense opportunities for extracting new forms of urban data through crowdsourcing. This book offers a comprehensive review of the methods that form the core of urban informatics from various kinds of urban remote sensing to new approaches to machine learning and statistical modelling. It provides a detailed technical introduction to the wide array of tools information scientists need to develop the key urban analytics that are fundamental to learning about the smart city, and it outlines ways in which these tools can be used to inform design and policy so that cities can become more efficient with a greater concern for environment and equity
A Machine Learning Enhanced Scheme for Intelligent Network Management
The versatile networking services bring about huge influence on daily living styles while the amount and diversity of services cause high complexity of network systems. The network scale and complexity grow with the increasing infrastructure apparatuses, networking function, networking slices, and underlying architecture evolution. The conventional way is manual administration to maintain the large and complex platform, which makes effective and insightful management troublesome. A feasible and promising scheme is to extract insightful information from largely produced network data. The goal of this thesis is to use learning-based algorithms inspired by machine learning communities to discover valuable knowledge from substantial network data, which directly promotes intelligent management and maintenance. In the thesis, the management and maintenance focus on two schemes: network anomalies detection and root causes localization; critical traffic resource control and optimization. Firstly, the abundant network data wrap up informative messages but its heterogeneity and perplexity make diagnosis challenging. For unstructured logs, abstract and formatted log templates are extracted to regulate log records. An in-depth analysis framework based on heterogeneous data is proposed in order to detect the occurrence of faults and anomalies. It employs representation learning methods to map unstructured data into numerical features, and fuses the extracted feature for network anomaly and fault detection. The representation learning makes use of word2vec-based embedding technologies for semantic expression. Next, the fault and anomaly detection solely unveils the occurrence of events while failing to figure out the root causes for useful administration so that the fault localization opens a gate to narrow down the source of systematic anomalies. The extracted features are formed as the anomaly degree coupled with an importance ranking method to highlight the locations of anomalies in network systems. Two types of ranking modes are instantiated by PageRank and operation errors for jointly highlighting latent issue of locations. Besides the fault and anomaly detection, network traffic engineering deals with network communication and computation resource to optimize data traffic transferring efficiency. Especially when network traffic are constrained with communication conditions, a pro-active path planning scheme is helpful for efficient traffic controlling actions. Then a learning-based traffic planning algorithm is proposed based on sequence-to-sequence model to discover hidden reasonable paths from abundant traffic history data over the Software Defined Network architecture. Finally, traffic engineering merely based on empirical data is likely to result in stale and sub-optimal solutions, even ending up with worse situations. A resilient mechanism is required to adapt network flows based on context into a dynamic environment. Thus, a reinforcement learning-based scheme is put forward for dynamic data forwarding considering network resource status, which explicitly presents a promising performance improvement. In the end, the proposed anomaly processing framework strengthens the analysis and diagnosis for network system administrators through synthesized fault detection and root cause localization. The learning-based traffic engineering stimulates networking flow management via experienced data and further shows a promising direction of flexible traffic adjustment for ever-changing environments