7 research outputs found
Prototype as Query for Few Shot Semantic Segmentation
Few-shot Semantic Segmentation (FSS) was proposed to segment unseen classes
in a query image, referring to only a few annotated examples named support
images. One of the characteristics of FSS is spatial inconsistency between
query and support targets, e.g., texture or appearance. This greatly challenges
the generalization ability of methods for FSS, which requires to effectively
exploit the dependency of the query image and the support examples. Most
existing methods abstracted support features into prototype vectors and
implemented the interaction with query features using cosine similarity or
feature concatenation. However, this simple interaction may not capture spatial
details in query features. To alleviate this limitation, a few methods utilized
all pixel-wise support information via computing the pixel-wise correlations
between paired query and support features implemented with the attention
mechanism of Transformer. These approaches suffer from heavy computation on the
dot-product attention between all pixels of support and query features. In this
paper, we propose a simple yet effective framework built upon Transformer
termed as ProtoFormer to fully capture spatial details in query features. It
views the abstracted prototype of the target class in support features as Query
and the query features as Key and Value embeddings, which are input to the
Transformer decoder. In this way, the spatial details can be better captured
and the semantic features of target class in the query image can be focused.
The output of the Transformer-based module can be viewed as semantic-aware
dynamic kernels to filter out the segmentation mask from the enriched query
features. Extensive experiments on PASCAL- and COCO- show that
our ProtoFormer significantly advances the state-of-the-art methods.Comment: under revie
Inter- and intra-uncertainty based feature aggregation model for semi-supervised histopathology image segmentation
Acquiring pixel-level annotations is often limited in applications such as
histology studies that require domain expertise. Various semi-supervised
learning approaches have been developed to work with limited ground truth
annotations, such as the popular teacher-student models. However, hierarchical
prediction uncertainty within the student model (intra-uncertainty) and image
prediction uncertainty (inter-uncertainty) have not been fully utilized by
existing methods. To address these issues, we first propose a novel inter- and
intra-uncertainty regularization method to measure and constrain both inter-
and intra-inconsistencies in the teacher-student architecture. We also propose
a new two-stage network with pseudo-mask guided feature aggregation (PG-FANet)
as the segmentation model. The two-stage structure complements with the
uncertainty regularization strategy to avoid introducing extra modules in
solving uncertainties and the aggregation mechanisms enable multi-scale and
multi-stage feature integration. Comprehensive experimental results over the
MoNuSeg and CRAG datasets show that our PG-FANet outperforms other
state-of-the-art methods and our semi-supervised learning framework yields
competitive performance with a limited amount of labeled data
Mutually enhanced multi-view information learning for segmentation of lung tumor in CT images
No description supplied.</p