17,801 research outputs found
When 3D Bounding-Box Meets SAM: Point Cloud Instance Segmentation with Weak-and-Noisy Supervision
Learning from bounding-boxes annotations has shown great potential in
weakly-supervised 3D point cloud instance segmentation. However, we observed
that existing methods would suffer severe performance degradation with
perturbed bounding box annotations. To tackle this issue, we propose a
complementary image prompt-induced weakly-supervised point cloud instance
segmentation (CIP-WPIS) method. CIP-WPIS leverages pretrained knowledge
embedded in the 2D foundation model SAM and 3D geometric prior to achieve
accurate point-wise instance labels from the bounding box annotations.
Specifically, CP-WPIS first selects image views in which 3D candidate points of
an instance are fully visible. Then, we generate complementary background and
foreground prompts from projections to obtain SAM 2D instance mask predictions.
According to these, we assign the confidence values to points indicating the
likelihood of points belonging to the instance. Furthermore, we utilize 3D
geometric homogeneity provided by superpoints to decide the final instance
label assignments. In this fashion, we achieve high-quality 3D point-wise
instance labels. Extensive experiments on both Scannet-v2 and S3DIS benchmarks
demonstrate that our method is robust against noisy 3D bounding-box annotations
and achieves state-of-the-art performance
Weakly-supervised Semantic Segmentation with Regularized Loss Hyperparameter Search
Weakly supervised segmentation signi cantly reduces user annotation e ort. Recently,
regularized loss was proposed for single object class segmentation under image-level weak
supervision. Regularized loss consists of several components. Each component, if used in
isolation, would lead to some trivial solution. However, a weighted combination of the loss
components introduces a balance between the individual biases. The weight of each component
in regularized loss is controlled by a hyperparameter. We propose an approach that
searches for regularized loss hyperparameters. The main idea is to set the most important
regularized loss component to a high weight while ensuring the other loss components are
set to weights just su ciently high to prevent the trivial solution favoured by the most
important component. Our approach results in a signi cantly improved performance over
prior work with xed hyperparameters and improves the state of the art in salient and
semantic image level supervised segmentation.
In addition to image level weak supervision, we propose a new approach for semantic
segmentation with weak supervision using bounding box annotations. Our new approach to
weak supervision from bounding boxes also makes use of hyperparameter search regularized
loss. Previous work on weak supervision from bounding boxes constructs pseudo-ground
truth by segmenting each box into the object and the background for each box independently
from all the other boxes in the dataset. We argue that the collection of boxes for
the same class naturally provides a dataset from which we can learn the appearance of that
object class. Learning a good appearance model, in turn, leads to a better segmentation of
each individual box. Thus for each class, we propose to train a segmentation CNN as from
the dataset consisting of the bounding boxes for that class using our proposed single object
approach. After we train these single-class CNNs, we apply them back to the training
bounding boxes to obtain object/background segmentations and merge them to construct
pseudo-ground truth. The obtained pseudo-ground truth is used for training a standard
segmentation CNN. We improve the state of the art on Pascal VOC 2012 benchmark in
bounding box weak supervision setting
Weakly- and Semi-Supervised Panoptic Segmentation
We present a weakly supervised model that jointly performs both semantic- and
instance-segmentation -- a particularly relevant problem given the substantial
cost of obtaining pixel-perfect annotation for these tasks. In contrast to many
popular instance segmentation approaches based on object detectors, our method
does not predict any overlapping instances. Moreover, we are able to segment
both "thing" and "stuff" classes, and thus explain all the pixels in the image.
"Thing" classes are weakly-supervised with bounding boxes, and "stuff" with
image-level tags. We obtain state-of-the-art results on Pascal VOC, for both
full and weak supervision (which achieves about 95% of fully-supervised
performance). Furthermore, we present the first weakly-supervised results on
Cityscapes for both semantic- and instance-segmentation. Finally, we use our
weakly supervised framework to analyse the relationship between annotation
quality and predictive performance, which is of interest to dataset creators.Comment: ECCV 2018. The first two authors contributed equall
Pixelwise Instance Segmentation with a Dynamically Instantiated Network
Semantic segmentation and object detection research have recently achieved
rapid progress. However, the former task has no notion of different instances
of the same object, and the latter operates at a coarse, bounding-box level. We
propose an Instance Segmentation system that produces a segmentation map where
each pixel is assigned an object class and instance identity label. Most
approaches adapt object detectors to produce segments instead of boxes. In
contrast, our method is based on an initial semantic segmentation module, which
feeds into an instance subnetwork. This subnetwork uses the initial
category-level segmentation, along with cues from the output of an object
detector, within an end-to-end CRF to predict instances. This part of our model
is dynamically instantiated to produce a variable number of instances per
image. Our end-to-end approach requires no post-processing and considers the
image holistically, instead of processing independent proposals. Therefore,
unlike some related work, a pixel cannot belong to multiple instances.
Furthermore, far more precise segmentations are achieved, as shown by our
state-of-the-art results (particularly at high IoU thresholds) on the Pascal
VOC and Cityscapes datasets.Comment: CVPR 201
- …