5 research outputs found

    People detection with omnidirectional cameras using aspatial grid ofdeep learning foveatic classifiers

    Full text link
    A novel deep-learning people detection algorithm using omnidirectional cameras is presented, which only requires point-based annotations, unlike most of the prominent works that require bounding box annotations. Thus, the effort of manually annotating the needed training databases is significantly reduced, allowing a faster system deployment. The algorithm is based on a novel deep neural network architecture that implements the concept of Grid of Spatial-Aware Classifiers, but allowing end-to-end training that improves the performance of the whole system. The designed algorithm satisfactorily handles the severe geometric distortions of the omnidirectional images, which typically degrades the performance of state-of-the-art detectors, without requiring any camera calibration. The algorithm has been evaluated in well-known omnidirectional image databases (PIROPO, BOMNI, and MW-18Mar) and compared with several works of the state of the art.This work has been partially supported by project PID2020115132RB (SARAOS) funded by MCIN/AEI/10.13039/501100011033 of the Spanish Government

    People counting using an overhead fisheye camera

    Full text link
    As climate change concerns grow, the reduction of energy consumption is seen as one of many potential solutions. In the US, a considerable amount of energy is wasted in commercial buildings due to sub-optimal heating, ventilation and air conditioning that operate with no knowledge of the occupancy level in various rooms and open areas. In this thesis, I develop an approach to passive occupancy estimation that does not require occupants to carry any type of beacon, but instead uses an overhead camera with fisheye lens (360 by 180 degree field of view). The difficulty with fisheye images is that occupants may appear not only in the upright position, but also upside-down, horizontally and diagonally, and thus algorithms developed for typical side-mounted, standard-lens cameras tend to fail. As the top-performing people detection algorithms today use deep learning, a logical step would be to develop and train a new neural-network model. However, there exist no large fisheye-image datasets with person annotations to facilitate training a new model. Therefore, I developed two people-counting methods that leverage YOLO (version 3), a state-of-the-art object detection method trained on standard datasets. In one approach, YOLO is applied to 24 rotated and highly-overlapping windows, and the results are post-processed to produce a people count. In the other approach, regions of interest are first extracted via background subtraction and only windows that include such regions are supplied to YOLO and post-processed. I carried out extensive experimental evaluation of both algorithms and showed their superior performance compared to a benchmark method
    corecore