42,297 research outputs found
Multi-utility Learning: Structured-output Learning with Multiple Annotation-specific Loss Functions
Structured-output learning is a challenging problem; particularly so because
of the difficulty in obtaining large datasets of fully labelled instances for
training. In this paper we try to overcome this difficulty by presenting a
multi-utility learning framework for structured prediction that can learn from
training instances with different forms of supervision. We propose a unified
technique for inferring the loss functions most suitable for quantifying the
consistency of solutions with the given weak annotation. We demonstrate the
effectiveness of our framework on the challenging semantic image segmentation
problem for which a wide variety of annotations can be used. For instance, the
popular training datasets for semantic segmentation are composed of images with
hard-to-generate full pixel labellings, as well as images with easy-to-obtain
weak annotations, such as bounding boxes around objects, or image-level labels
that specify which object categories are present in an image. Experimental
evaluation shows that the use of annotation-specific loss functions
dramatically improves segmentation accuracy compared to the baseline system
where only one type of weak annotation is used
Articulated Pose Estimation Using Hierarchical Exemplar-Based Models
Exemplar-based models have achieved great success on localizing the parts of
semi-rigid objects. However, their efficacy on highly articulated objects such
as humans is yet to be explored. Inspired by hierarchical object representation
and recent application of Deep Convolutional Neural Networks (DCNNs) on human
pose estimation, we propose a novel formulation that incorporates both
hierarchical exemplar-based models and DCNNs in the spatial terms.
Specifically, we obtain more expressive spatial models by assuming independence
between exemplars at different levels in the hierarchy; we also obtain stronger
spatial constraints by inferring the spatial relations between parts at the
same level. As our method strikes a good balance between expressiveness and
strength of spatial models, it is both effective and generalizable, achieving
state-of-the-art results on different benchmarks: Leeds Sports Dataset and
CUB-200-2011.Comment: 8 pages, 6 figure
NLSC: Unrestricted Natural Language-based Service Composition through Sentence Embeddings
Current approaches for service composition (assemblies of atomic services)
require developers to use: (a) domain-specific semantics to formalize services
that restrict the vocabulary for their descriptions, and (b) translation
mechanisms for service retrieval to convert unstructured user requests to
strongly-typed semantic representations. In our work, we argue that effort to
developing service descriptions, request translations, and matching mechanisms
could be reduced using unrestricted natural language; allowing both: (1)
end-users to intuitively express their needs using natural language, and (2)
service developers to develop services without relying on syntactic/semantic
description languages. Although there are some natural language-based service
composition approaches, they restrict service retrieval to syntactic/semantic
matching. With recent developments in Machine learning and Natural Language
Processing, we motivate the use of Sentence Embeddings by leveraging richer
semantic representations of sentences for service description, matching and
retrieval. Experimental results show that service composition development
effort may be reduced by more than 44\% while keeping a high precision/recall
when matching high-level user requests with low-level service method
invocations.Comment: This paper will appear on SCC'19 (IEEE International Conference on
Services Computing) on July 1
- …
