2 research outputs found
Occlusion-shared and Feature-separated Network for Occlusion Relationship Reasoning
Occlusion relationship reasoning demands closed contour to express the
object, and orientation of each contour pixel to describe the order
relationship between objects. Current CNN-based methods neglect two critical
issues of the task: (1) simultaneous existence of the relevance and distinction
for the two elements, i.e, occlusion edge and occlusion orientation; and (2)
inadequate exploration to the orientation features. For the reasons above, we
propose the Occlusion-shared and Feature-separated Network (OFNet). On one
hand, considering the relevance between edge and orientation, two sub-networks
are designed to share the occlusion cue. On the other hand, the whole network
is split into two paths to learn the high-level semantic features separately.
Moreover, a contextual feature for orientation prediction is extracted, which
represents the bilateral cue of the foreground and background areas. The
bilateral cue is then fused with the occlusion cue to precisely locate the
object regions. Finally, a stripe convolution is designed to further aggregate
features from surrounding scenes of the occlusion edge. The proposed OFNet
remarkably advances the state-of-the-art approaches on PIOD and BSDS ownership
dataset. The source code is available at https://github.com/buptlr/OFNet.Comment: Accepted by ICCV 2019. Code and pretrained model are available at
https://github.com/buptlr/OFNe
Robust Instance Segmentation through Reasoning about Multi-Object Occlusion
Analyzing complex scenes with Deep Neural Networks is a challenging task,
particularly when images contain multiple objects that partially occlude each
other. Existing approaches to image analysis mostly process objects
independently and do not take into account the relative occlusion of nearby
objects. In this paper, we propose a deep network for multi-object instance
segmentation that is robust to occlusion and can be trained from bounding box
supervision only. Our work builds on Compositional Networks, which learn a
generative model of neural feature activations to locate occluders and to
classify objects based on their non-occluded parts. We extend their generative
model to include multiple objects and introduce a framework for efficient
inference in challenging occlusion scenarios. In particular, we obtain
feed-forward predictions of the object classes and their instance and occluder
segmentations. We introduce an Occlusion Reasoning Module (ORM) that locates
erroneous segmentations and estimates the occlusion order to correct them. The
improved segmentation masks are, in turn, integrated into the network in a
top-down manner to improve the image classification. Our experiments on the
KITTI INStance dataset (KINS) and a synthetic occlusion dataset demonstrate the
effectiveness and robustness of our model at multi-object instance segmentation
under occlusion. Code is publically available at
https://github.com/XD7479/Multi-Object-Occlusion.Comment: Accepted by CVPR 202