37,546 research outputs found
Measuring the Accuracy of Object Detectors and Trackers
The accuracy of object detectors and trackers is most commonly evaluated by
the Intersection over Union (IoU) criterion. To date, most approaches are
restricted to axis-aligned or oriented boxes and, as a consequence, many
datasets are only labeled with boxes. Nevertheless, axis-aligned or oriented
boxes cannot accurately capture an object's shape. To address this, a number of
densely segmented datasets has started to emerge in both the object detection
and the object tracking communities. However, evaluating the accuracy of object
detectors and trackers that are restricted to boxes on densely segmented data
is not straightforward. To close this gap, we introduce the relative
Intersection over Union (rIoU) accuracy measure. The measure normalizes the IoU
with the optimal box for the segmentation to generate an accuracy measure that
ranges between 0 and 1 and allows a more precise measurement of accuracies.
Furthermore, it enables an efficient and easy way to understand scenes and the
strengths and weaknesses of an object detection or tracking approach. We
display how the new measure can be efficiently calculated and present an
easy-to-use evaluation framework. The framework is tested on the DAVIS and the
VOT2016 segmentations and has been made available to the community.Comment: 10 pages, 7 Figure
Going Further with Point Pair Features
Point Pair Features is a widely used method to detect 3D objects in point
clouds, however they are prone to fail in presence of sensor noise and
background clutter. We introduce novel sampling and voting schemes that
significantly reduces the influence of clutter and sensor noise. Our
experiments show that with our improvements, PPFs become competitive against
state-of-the-art methods as it outperforms them on several objects from
challenging benchmarks, at a low computational cost.Comment: Corrected post-print of manuscript accepted to the European
Conference on Computer Vision (ECCV) 2016;
https://link.springer.com/chapter/10.1007/978-3-319-46487-9_5
Efficient Algorithms for Distributed Detection of Holes and Boundaries in Wireless Networks
We propose two novel algorithms for distributed and location-free boundary
recognition in wireless sensor networks. Both approaches enable a node to
decide autonomously whether it is a boundary node, based solely on connectivity
information of a small neighborhood. This makes our algorithms highly
applicable for dynamic networks where nodes can move or become inoperative.
We compare our algorithms qualitatively and quantitatively with several
previous approaches. In extensive simulations, we consider various models and
scenarios. Although our algorithms use less information than most other
approaches, they produce significantly better results. They are very robust
against variations in node degree and do not rely on simplified assumptions of
the communication model. Moreover, they are much easier to implement on real
sensor nodes than most existing approaches.Comment: extended version of accepted submission to SEA 201
- …