22,107 research outputs found
What makes for effective detection proposals?
Current top performing object detectors employ detection proposals to guide
the search for objects, thereby avoiding exhaustive sliding window search
across images. Despite the popularity and widespread use of detection
proposals, it is unclear which trade-offs are made when using them during
object detection. We provide an in-depth analysis of twelve proposal methods
along with four baselines regarding proposal repeatability, ground truth
annotation recall on PASCAL, ImageNet, and MS COCO, and their impact on DPM,
R-CNN, and Fast R-CNN detection performance. Our analysis shows that for object
detection improving proposal localisation accuracy is as important as improving
recall. We introduce a novel metric, the average recall (AR), which rewards
both high recall and good localisation and correlates surprisingly well with
detection performance. Our findings show common strengths and weaknesses of
existing methods, and provide insights and metrics for selecting and tuning
proposal methods.Comment: TPAMI final version, duplicate proposals removed in experiment
S4Net: Single Stage Salient-Instance Segmentation
We consider an interesting problem-salient instance segmentation in this
paper. Other than producing bounding boxes, our network also outputs
high-quality instance-level segments. Taking into account the
category-independent property of each target, we design a single stage salient
instance segmentation framework, with a novel segmentation branch. Our new
branch regards not only local context inside each detection window but also its
surrounding context, enabling us to distinguish the instances in the same scope
even with obstruction. Our network is end-to-end trainable and runs at a fast
speed (40 fps when processing an image with resolution 320x320). We evaluate
our approach on a publicly available benchmark and show that it outperforms
other alternative solutions. We also provide a thorough analysis of the design
choices to help readers better understand the functions of each part of our
network. The source code can be found at
\url{https://github.com/RuochenFan/S4Net}
Online Object Tracking with Proposal Selection
Tracking-by-detection approaches are some of the most successful object
trackers in recent years. Their success is largely determined by the detector
model they learn initially and then update over time. However, under
challenging conditions where an object can undergo transformations, e.g.,
severe rotation, these methods are found to be lacking. In this paper, we
address this problem by formulating it as a proposal selection task and making
two contributions. The first one is introducing novel proposals estimated from
the geometric transformations undergone by the object, and building a rich
candidate set for predicting the object location. The second one is devising a
novel selection strategy using multiple cues, i.e., detection score and
edgeness score computed from state-of-the-art object edges and motion
boundaries. We extensively evaluate our approach on the visual object tracking
2014 challenge and online tracking benchmark datasets, and show the best
performance.Comment: ICCV 201
Object Detection based on Region Decomposition and Assembly
Region-based object detection infers object regions for one or more
categories in an image. Due to the recent advances in deep learning and region
proposal methods, object detectors based on convolutional neural networks
(CNNs) have been flourishing and provided the promising detection results.
However, the detection accuracy is degraded often because of the low
discriminability of object CNN features caused by occlusions and inaccurate
region proposals. In this paper, we therefore propose a region decomposition
and assembly detector (R-DAD) for more accurate object detection.
In the proposed R-DAD, we first decompose an object region into multiple
small regions. To capture an entire appearance and part details of the object
jointly, we extract CNN features within the whole object region and decomposed
regions. We then learn the semantic relations between the object and its parts
by combining the multi-region features stage by stage with region assembly
blocks, and use the combined and high-level semantic features for the object
classification and localization. In addition, for more accurate region
proposals, we propose a multi-scale proposal layer that can generate object
proposals of various scales. We integrate the R-DAD into several feature
extractors, and prove the distinct performance improvement on PASCAL07/12 and
MSCOCO18 compared to the recent convolutional detectors.Comment: Accepted to 2019 AAAI Conference on Artificial Intelligence (AAAI
DeepProposals: Hunting Objects and Actions by Cascading Deep Convolutional Layers
In this paper, a new method for generating object and action proposals in
images and videos is proposed. It builds on activations of different
convolutional layers of a pretrained CNN, combining the localization accuracy
of the early layers with the high informative-ness (and hence recall) of the
later layers. To this end, we build an inverse cascade that, going backward
from the later to the earlier convolutional layers of the CNN, selects the most
promising locations and refines them in a coarse-to-fine manner. The method is
efficient, because i) it re-uses the same features extracted for detection, ii)
it aggregates features using integral images, and iii) it avoids a dense
evaluation of the proposals thanks to the use of the inverse coarse-to-fine
cascade. The method is also accurate. We show that our DeepProposals outperform
most of the previously proposed object proposal and action proposal approaches
and, when plugged into a CNN-based object detector, produce state-of-the-art
detection performance.Comment: 15 page
- …