229 research outputs found
Real-time image streaming over a low-bandwidth wireless camera network
In this paper we describe the recent development of a low-bandwidth wireless camera sensor network. We propose a simple, yet effective, network architecture which allows multiple cameras to be connected to the network and synchronize their communication schedules. Image compression of greater than 90% is performed at each node running on a local DSP coprocessor, resulting in nodes using 1/8th the energy compared to streaming uncompressed images. We briefly introduce the Fleck wireless node and the DSP/camera sensor, and then outline the network architecture and compression algorithm. The system is able to stream color QVGA images over the network to a base station at up to 2 frames per second. Ă© 2007 IEEE
Secure Chaotic Map Based Block Cryptosystem with Application to Camera Sensor Networks
Recently, Wang et al. presented an efficient logistic map based block encryption system. The encryption system employs feedback ciphertext to achieve plaintext dependence of sub-keys. Unfortunately, we discovered that their scheme is unable to withstand key stream attack. To improve its security, this paper proposes a novel chaotic map based block cryptosystem. At the same time, a secure architecture for camera sensor network is constructed. The network comprises a set of inexpensive camera sensors to capture the images, a sink node equipped with sufficient computation and storage capabilities and a data processing server. The transmission security between the sink node and the server is gained by utilizing the improved cipher. Both theoretical analysis and simulation results indicate that the improved algorithm can overcome the flaws and maintain all the merits of the original cryptosystem. In addition, computational costs and efficiency of the proposed scheme are encouraging for the practical implementation in the real environment as well as camera sensor network
Distributed Calibration of a Camera Sensor Network
The thesis proposes algorithms to perform distributed calibration of a camera sensor network. Unlike other calibration techniques which involve non-linear computations, we propose simple techniques which can be implemented on low power processors. The concept of pin-hole camera model and generation of a reference coordinate system that are used in this thesis, were inspired from some existing work. A cooperative friendly target equipped with a dead reckoning position sensor, moves around in the environment and communicates its coordinate information at three distinct locations in the field of view of a camera sensor node. The sensor node uses the concept of slope angle of a straight line to solve for its location and orientation. This method does not require any initial guesses to calculate the external parameters of a camera. The correctness of the method was verified experimentally and its scalability for a large scale sensor network was validated through simulation.School of Electrical & Computer Engineerin
A Reaction-Diffusion-Based Coding Rate Control Mechanism for Camera Sensor Networks
A wireless camera sensor network is useful for surveillance and monitoring for its visibility and easy deployment. However, it suffers from the limited capacity of wireless communication and a network is easily overflown with a considerable amount of video traffic. In this paper, we propose an autonomous video coding rate control mechanism where each camera sensor node can autonomously determine its coding rate in accordance with the location and velocity of target objects. For this purpose, we adopted a biological model, i.e., reaction-diffusion model, inspired by the similarity of biological spatial patterns and the spatial distribution of video coding rate. Through simulation and practical experiments, we verify the effectiveness of our proposal
Vision Graph Construction in Wireless Multimedia Sensor Networks
In Wireless multimedia sensor networks (WMSNs), the camera nodes connected in the vision graph share overlapped field of views (FOVs) and they depend on the densely deployed relay nodes in the communication network graph to communicate with each other. Given a uniformly deployed camera sensor network with relay nodes, the problem is to find the number of hops for the vision-graph-neighbor-searching messages to construct the vision graph in an energy efficient way. In this paper, mathematical models are developed to analyze the FOV overlap of the camera nodes and the multi-hop communications in two dimensional topologies, which are utilized to analyze the relation between vision graph construction and maximum hop count. In addition, simulations are conducted to verify the models
On Achieving Diversity in the Presence of Outliers in Participatory Camera Sensor Networks
This paper addresses the problem of collection and
delivery of a representative subset of pictures, in participatory camera networks, to maximize coverage when a significant portion of the pictures may be redundant or irrelevant. Consider, for example, a rescue mission where volunteers and survivors of a large-scale disaster scout a wide area to capture pictures of
damage in distressed neighborhoods, using handheld cameras, and report them to a rescue station. In this participatory camera network, a significant amount of pictures may be redundant (i.e., similar pictures may be reported by many) or irrelevant (i.e., may
not document an event of interest). Given this pool of pictures, we aim to build a protocol to store and deliver a smaller subset of pictures, among all those taken, that minimizes redundancy and eliminates irrelevant objects and outliers. While previous work addressed removal of redundancy alone, doing so in the presence of outliers is tricky, because outliers, by their very nature, are different from other objects, causing redundancy minimizing algorithms to favor their inclusion, which is at odds with the goal of finding a representative subset. To eliminate both outliers and redundancy at the same time, two seemingly opposite objectives must be met together. The contribution of this
paper lies in a new prioritization technique (and its in-network
implementation) that minimizes redundancy among delivered
pictures, while also reducing outliers.unpublishedis peer reviewe
Distributed Feature Extraction Using Cloud Computing Resources
The need to expand the computational resources in a massive surveillance network is clear but traditional means of purchasing new equipment for short-term tasks every year is wasteful. In this work I will provide evidence in support of utilizing a cloud computing infrastructure to perform computationally intensive feature extraction tasks on data streams. Efficient off-loading of computational tasks to cloud resources will require a minimization of the time needed to expand the cloud resources, an efficient model of communication and a study of the interplay between the in-network computational resources and remote resources in the cloud. This report provides strong evidence that the use of cloud computing resources in a near real-time distributed sensor network surveillance system, ASAP, is feasible. A face detection web service operating on an Amazon EC2 instance is shown to provide processing of 10-15 frames per second.Umakishore Ramachandran - Faculty Mentor ; Rajnish Kumar - Committee Member/Second Reade
ITR/SY: a distributed programming infrastructure for integrating smart sensors
Issued as final reportNational Science Foundation (U.S.
A new Measure for Optimization of Field Sensor Network with Application to LiDAR
This thesis proposes a solution to the problem of modeling and optimizing the field sensor network in terms of the coverage performance. The term field sensor is referred to a class of sensors which can detect the regions in 2D/3D spaces through non-contact measurements. The most widely used field sensors include cameras, LiDAR, ultrasonic sensor, and RADAR, etc. The key challenge in the applications of field sensor networks, such as area coverage, is to develop an effective performance measure, which has to involve both sensor and environment parameters. The nature of space distribution in the case of the field sensor incurs a great deal of difficulties for such development and, hence, poses it as a very interesting research problem. Therefore, to tackle this problem, several attempts have been made in the literature. However, they have failed to address a comprehensive and applicable approach to distinctive types of field sensors (in 3D), as only coverage of a particular sensor is usually addressed at the time. In addition, no coverage model has been proposed yet for some types of field sensors such as LiDAR sensors. In this dissertation, a coverage model is obtained for the field sensors based on the transformation of sensor and task parameters into the sensor geometric model. By providing a mathematical description of the sensorâs sensing region, a performance measure is introduced which characterizes the closeness between a single sensor and target configurations. In this regard, the first contribution is developing an Infinity norm based measure which describes the target distance to the closure of the sensing region expressed by an area-based approach. The second contribution can be geometrically interpreted as mapping the sensorâs sensing region to an n-ball using a homeomorphism map and developing a performance measure. The third contribution is introducing the measurement principle and establishing the coverage model for the class of solid-state (flash) LiDAR sensors. The fourth contribution is point density analysis and developing the coverage model for the class of mechanical (prism rotating mechanism) LiDAR sensors. Finally, the effectiveness of the proposed coverage model is illustrated by simulations, experiments, and comparisons is carried out throughout the dissertation. This coverage model is a powerful tool as it applies to the variety of field sensors
- âŠ