5,281 research outputs found
Dynamic Reconfiguration in Camera Networks: A Short Survey
There is a clear trend in camera networks towards enhanced functionality and flexibility, and a fixed static deployment is typically not sufficient to fulfill these increased requirements. Dynamic network reconfiguration helps to optimize the network performance to the currently required specific tasks while considering the available resources. Although several reconfiguration methods have been recently proposed, e.g., for maximizing the global scene coverage or maximizing the image quality of specific targets, there is a lack of a general framework highlighting the key components shared by all these systems. In this paper we propose a reference framework for network reconfiguration and present a short survey of some of the most relevant state-of-the-art works in this field, showing how they can be reformulated in our framework. Finally we discuss the main open research challenges in camera network reconfiguration
Distributed Object Tracking Using a Cluster-Based Kalman Filter in Wireless Camera Networks
Local data aggregation is an effective means to save sensor node energy and prolong the lifespan of wireless sensor networks. However, when a sensor network is used to track moving objects, the task of local data aggregation in the network presents a new set of challenges, such as the necessity to estimate, usually in real time, the constantly changing state of the target based on information acquired by the nodes at different time instants. To address these issues, we propose a distributed object tracking system which employs a cluster-based Kalman filter in a network of wireless cameras. When a target is detected, cameras that can observe the same target interact with one another to form a cluster and elect a cluster head. Local measurements of the target acquired by members of the cluster are sent to the cluster head, which then estimates the target position via Kalman filtering and periodically transmits this information to a base station. The underlying clustering protocol allows the current state and uncertainty of the target position to be easily handed off among clusters as the object is being tracked. This allows Kalman filter-based object tracking to be carried out in a distributed manner. An extended Kalman filter is necessary since measurements acquired by the cameras are related to the actual position of the target by nonlinear transformations. In addition, in order to take into consideration the time uncertainty in the measurements acquired by the different cameras, it is necessary to introduce nonlinearity in the system dynamics. Our object tracking protocol requires the transmission of significantly fewer messages than a centralized tracker that naively transmits all of the local measurements to the base station. It is also more accurate than a decentralized tracker that employs linear interpolation for local data aggregation. Besides, the protocol is able to perform real-time estimation because our implementation takes into consideration the sparsit- - y of the matrices involved in the problem. The experimental results show that our distributed object tracking protocol is able to achieve tracking accuracy comparable to the centralized tracking method, while requiring a significantly smaller number of message transmissions in the network
Collaborative Solutions to Visual Sensor Networks
Visual sensor networks (VSNs) merge computer vision, image processing and wireless sensor network disciplines to solve problems in multi-camera applications in large surveillance areas. Although potentially powerful, VSNs also present unique challenges that could hinder their practical deployment because of the unique camera features including the extremely higher data rate, the directional sensing characteristics, and the existence of visual occlusions.
In this dissertation, we first present a collaborative approach for target localization in VSNs. Traditionally; the problem is solved by localizing targets at the intersections of the back-projected 2D cones of each target. However, the existence of visual occlusions among targets would generate many false alarms. Instead of resolving the uncertainty about target existence at the intersections, we identify and study the non-occupied areas in 2D cones and generate the so-called certainty map of targets non-existence. We also propose distributed integration of local certainty maps by following a dynamic itinerary where the entire map is progressively clarified.
The accuracy of target localization is affected by the existence of faulty nodes in VSNs. Therefore, we present the design of a fault-tolerant localization algorithm that would not only accurately localize targets but also detect the faults in camera orientations, tolerate these errors and further correct them before they cascade. Based on the locations of detected targets in the fault-tolerated final certainty map, we construct a generative image model that estimates the camera orientations, detect inaccuracies and correct them.
In order to ensure the required visual coverage to accurately localize targets or tolerate the faulty nodes, we need to calculate the coverage before deploying sensors. Therefore, we derive the closed-form solution for the coverage estimation based on the certainty-based detection model that takes directional sensing of cameras and existence of visual occlusions into account.
The effectiveness of the proposed collaborative and fault-tolerant target localization algorithms in localization accuracy as well as fault detection and correction performance has been validated through the results obtained from both simulation and real experiments. In addition, conducted simulation shows extreme consistency with results from theoretical closed-form solution for visual coverage estimation, especially when considering the boundary effect
Self-localizing Smart Cameras and Their Applications
As the prices of cameras and computing elements continue to fall, it
has become increasingly attractive to consider the deployment of
smart camera networks. These networks would be composed of small,
networked computers equipped with inexpensive image sensors. Such
networks could be employed in a wide range of applications including
surveillance, robotics and 3D scene reconstruction.
One critical problem that must be addressed before such systems can
be deployed effectively is the issue of localization. That is, in
order to take full advantage of the images gathered from multiple
vantage points it is helpful to know how the cameras in the scene
are positioned and oriented with respect to each other. To address
the localization problem we have proposed a novel approach to
localizing networks of embedded cameras and sensors. In this scheme
the cameras and the nodes are equipped with controllable light
sources (either visible or infrared) which are used for
signaling. Each camera node can then automatically determine the
bearing to all the nodes that are visible from its vantage point. By
fusing these measurements with the measurements obtained from
onboard accelerometers, the camera nodes are able to determine the
relative positions and orientations of other nodes in the network.
This localization technology can serve as a basic capability on
which higher level applications can be built. The method could be
used to automatically survey the locations of sensors of interest,
to implement distributed surveillance systems or to analyze the
structure of a scene based on the images obtained from multiple
registered vantage points. It also provides a mechanism for
integrating the imagery obtained from the cameras with the
measurements obtained from distributed sensors.
We have successfully used our custom made self localizing smart
camera networks to implement a novel decentralized target tracking
algorithm, create an ad-hoc range finder and localize the components
of a self assembling modular robot
Power Management in Sensing Subsystem of Wireless Multimedia Sensor Networks
A wireless sensor network consists of sensor
nodes deployed over a geographical area for
monitoring physical phenomena like temperature, humidity, vibrations, seismic events, and so on. Typically, a sensor node is a tiny device that includes three basic components: a sensing subsystem for data acquisition from the physical surrounding environment, a processing subsystem for local data processing and storage, and a wireless communication subsystem for data transmission. In addition, a power source supplies the energy needed by the device to perform the programmed task. This power source often consists of a battery with a limited energy budget. In addition, it is usually impossible or inconvenient to recharge the battery, because nodes are deployed in a hostile or unpractical environment.
On the other hand, the sensor network should
have a lifetime long enough to fulfill the
application requirements. Accordingly, energy conservation in nodes and maximization of network lifetime are commonly recognized as a key challenge in the design and implementation of WSNs. Experimental measurements have shown that generally data transmission is very expensive in terms of energy consumption, while data processing consumes significantly less (Raghunathan et al., 2002). The energy cost of transmitting a single bit of information is approximately the same as that needed for processing a thousand operations in a typical sensor node (Pottie &
Kaiser, 2000). The energy consumption of the
sensing subsystem depends on the specific
sensor type. In some cases of scalar sensors,
it is negligible with respect to the energy
consumed by the processing and, above all, the communication subsystems. In other cases, the energy expenditure for data sensing may be comparable to, or even greater (in the case of multimedia sensing) than the energy needed for data transmission. In general, energy-saving
techniques focus on two subsystems: the communication subsystem (i.e., energy management is taken into account in the operations of each single node, as well as in the design of networking protocols), and the sensing subsystem (i.e., techniques are used to reduce the amount or frequency of energy-expensive samples).Postprint (published version
DESIGN FRAMEWORK FOR INTERNET OF THINGS BASED NEXT GENERATION VIDEO SURVEILLANCE
Modern artificial intelligence and machine learning opens up new era towards video
surveillance system. Next generation video surveillance in Internet of Things (IoT) environment is
an emerging research area because of high bandwidth, big-data generation, resource constraint
video surveillance node, high energy consumption for real time applications. In this thesis, various
opportunities and functional requirements that next generation video surveillance system should
achieve with the power of video analytics, artificial intelligence and machine learning are
discussed. This thesis also proposes a new video surveillance system architecture introducing fog
computing towards IoT based system and contributes the facilities and benefits of proposed system
which can meet the forthcoming requirements of surveillance. Different challenges and issues
faced for video surveillance in IoT environment and evaluate fog-cloud integrated architecture to
penetrate and eliminate those issues.
The focus of this thesis is to evaluate the IoT based video surveillance system. To this end,
two case studies were performed to penetrate values towards energy and bandwidth efficient video
surveillance system. In one case study, an IoT-based power efficient color frame transmission and
generation algorithm for video surveillance application is presented. The conventional way is to
transmit all R, G and B components of all frames. Using proposed technique, instead of sending
all components, first one color frame is sent followed by a series of gray-scale frames. After a
certain number of gray-scale frames, another color frame is sent followed by the same number of
gray-scale frames. This process is repeated for video surveillance system. In the decoder, color
information is formulated from the color frame and then used to colorize the gray-scale frames. In
another case study, a bandwidth efficient and low complexity frame reproduction technique that is
also applicable in IoT based video surveillance application is presented. Using the second
technique, only the pixel intensity that differs heavily comparing to previous frame’s
corresponding pixel is sent. If the pixel intensity is similar or near similar comparing to the
previous frame, the information is not transferred. With this objective, the bit stream is created for
every frame with a predefined protocol. In cloud side, the frame information can be reproduced by
implementing the reverse protocol from the bit stream.
Experimental results of the two case studies show that the IoT-based proposed approach
gives better results than traditional techniques in terms of both energy efficiency and quality of the video, and therefore, can enable sensor nodes in IoT to perform more operations with energy
constraints
Feature-based calibration of distributed smart stereo camera networks
A distributed smart camera network is a collective of vision-capable devices with enough processing power to execute algorithms for collaborative vision tasks. A true 3D sensing network applies to a broad range of applications, and local stereo vision capabilities at each node offer the potential for a particularly robust implementation. A novel spatial calibration method for such a network is presented, which obtains pose estimates suitable for collaborative 3D vision in a distributed fashion using two stages of registration on robust 3D features. The method is first described in a general, modular sense, assuming some ideal vision and registration algorithms. Then, existing algorithms are selected for a practical implementation. The method is designed independently of networking details, making only a few basic assumptions about the underlying network\u27s capabilities. Experiments using both software simulations and physical devices are designed and executed to demonstrate performance
- …