6 research outputs found

    A sparsity-driven approach to multi-camera tracking in visual sensor networks

    Get PDF
    In this paper, a sparsity-driven approach is presented for multi-camera tracking in visual sensor networks (VSNs). VSNs consist of image sensors, embedded processors and wireless transceivers which are powered by batteries. Since the energy and bandwidth resources are limited, setting up a tracking system in VSNs is a challenging problem. Motivated by the goal of tracking in a bandwidth-constrained environment, we present a sparsity-driven method to compress the features extracted by the camera nodes, which are then transmitted across the network for distributed inference. We have designed special overcomplete dictionaries that match the structure of the features, leading to very parsimonious yet accurate representations. We have tested our method in indoor and outdoor people tracking scenarios. Our experimental results demonstrate how our approach leads to communication savings without significant loss in tracking performance

    Multi-View Face Recognition From Single RGBD Models of the Faces

    Get PDF
    This work takes important steps towards solving the following problem of current interest: Assuming that each individual in a population can be modeled by a single frontal RGBD face image, is it possible to carry out face recognition for such a population using multiple 2D images captured from arbitrary viewpoints? Although the general problem as stated above is extremely challenging, it encompasses subproblems that can be addressed today. The subproblems addressed in this work relate to: (1) Generating a large set of viewpoint dependent face images from a single RGBD frontal image for each individual; (2) using hierarchical approaches based on view-partitioned subspaces to represent the training data; and (3) based on these hierarchical approaches, using a weighted voting algorithm to integrate the evidence collected from multiple images of the same face as recorded from different viewpoints. We evaluate our methods on three datasets: a dataset of 10 people that we created and two publicly available datasets which include a total of 48 people. In addition to providing important insights into the nature of this problem, our results show that we are able to successfully recognize faces with accuracies of 95% or higher, outperforming existing state-of-the-art face recognition approaches based on deep convolutional neural networks

    A Sparsity-Driven Approach to Multi-camera Tracking in Visual Sensor Networks

    Get PDF
    International audienceIn this paper, a sparsity-driven approach is presented for multi-camera tracking in visual sensor networks (VSNs). VSNs consist of image sensors, embedded processors and wireless transceivers which are powered by batteries. Since the energy and bandwidth resources are limited, setting up a tracking system in VSNs is a challenging problem. Motivated by the goal of tracking in a bandwidth-constrained environment , we present a sparsity-driven method to compress the features extracted by the camera nodes, which are then transmitted across the network for distributed inference. We have designed special overcomplete dictionaries that match the structure of the features, leading to very parsimonious yet accurate representations. We have tested our method in indoor and outdoor people tracking scenarios. Our experimental results demonstrate how our approach leads to communication savings without significant loss in tracking performance

    The costs of fusion in smart camera networks

    Get PDF
    ABSTRACT The choice of the most suitable fusion scheme for smart camera networks depends on the application as well as on the available computational and communication resources. In this paper we discuss and compare the resource requirements of five fusion schemes, namely centralised fusion, flooding, consensus, token passing and dynamic clustering. The Extended Information Filter is applied to each fusion scheme to perform target tracking. Token passing and dynamic clustering involve negotiation among viewing nodes (cameras observing the same target) to decide which node should perform the fusion process whereas flooding and consensus do not include this negotiation. Negotiation helps limiting the number of participating cameras and reduces the required resources for the fusion process itself but requires additional communication. Consensus has the highest communication and computation costs but it is the only scheme that can be applied when not all viewing nodes are connected directly and routing tables are not available

    Cluster-Based Distributed Face Tracking in Camera Networks

    Get PDF
    Abstract—In this paper, we present a distributed multicamera face tracking system suitable for large wired camera networks. Unlike previous multicamera face tracking systems, our system does not require a central server to coordinate the entire tracking effort. Instead, an efficient camera clustering protocol is used to dynamically form groups of cameras for in-network tracking of individual faces. The clustering protocol includes cluster propagation mechanisms that allow the computational load of face tracking to be transferred to different cameras as the target objects move. Furthermore, the dynamic election of cluster leaders provides robustness against system failures. Our experimental results show that our cluster-based distributed face tracker is capable of accurately tracking multiple faces in real-time. The overall performance of the distributed system is comparable to that of a centralized face tracker, while presenting the advantages of scalability and robustness. Index Terms—Camera networks, distributed tracking, face tracking, object detection. I

    Cluster-based distributed face tracking in camera networks

    No full text
    Distributed communication networks offer the advantages of scalability and not having a single point of failure. In distributed camera networks, each camera is equipped with an on-board processor capable of analyzing its images locally thus allowing most communication to take place between nearby cameras, with only summarized information being transmitted to a central server. Based on the wireless sensor cluster-based protocol developed by Medeiros et al. (2008), this thesis presents a multi-person tracking framework that represents individual observations of face pose from multiple cameras in a shared world coordinate system, allowing observations to be compared and integrated without reference to individual camera parameters. We show that the cluster-based protocol is useful for multi-object tracking not only in wireless camera networks, but also in wired camera networks. From a theoretical perspective, we show how the cluster-based protocol addresses computer vision challenges in wired networks using the same mechanisms that address communication challenges in wireless networks. From an empirical perspective, we demonstrate distributed tracking of people in real time in a twelve-camera wired network. This dissertation also proposes a benchmark for unconstrained face recognition based on human familiar face recognition. Face recognition holds the potential for greatly improving person tracking. However, current face recognition algorithms do not work reliably in the unconstrained images captured by camera networks. Our benchmark would allow future face recognition algorithms to be tested on the sorts of challenging images captured in a camera network and, at the same time, allow a comparison of the performance of these algorithms to unconstrained face recognition by humans
    corecore