1,423 research outputs found
Deep Convolutional Ranking for Multilabel Image Annotation
Multilabel image annotation is one of the most important challenges in
computer vision with many real-world applications. While existing work usually
use conventional visual features for multilabel annotation, features based on
Deep Neural Networks have shown potential to significantly boost performance.
In this work, we propose to leverage the advantage of such features and analyze
key components that lead to better performances. Specifically, we show that a
significant performance gain could be obtained by combining convolutional
architectures with approximate top- ranking objectives, as thye naturally
fit the multilabel tagging problem. Our experiments on the NUS-WIDE dataset
outperforms the conventional visual features by about 10%, obtaining the best
reported performance in the literature
Semi-Supervised Sparse Coding
Sparse coding approximates the data sample as a sparse linear combination of
some basic codewords and uses the sparse codes as new presentations. In this
paper, we investigate learning discriminative sparse codes by sparse coding in
a semi-supervised manner, where only a few training samples are labeled. By
using the manifold structure spanned by the data set of both labeled and
unlabeled samples and the constraints provided by the labels of the labeled
samples, we learn the variable class labels for all the samples. Furthermore,
to improve the discriminative ability of the learned sparse codes, we assume
that the class labels could be predicted from the sparse codes directly using a
linear classifier. By solving the codebook, sparse codes, class labels and
classifier parameters simultaneously in a unified objective function, we
develop a semi-supervised sparse coding algorithm. Experiments on two
real-world pattern recognition problems demonstrate the advantage of the
proposed methods over supervised sparse coding methods on partially labeled
data sets
SPar: A DSL for High-Level and Productive Stream Parallelism
This paper introduces SPar, an internal C++ Domain-Specific Language (DSL) that supports the development of classic stream parallel applications. The DSL uses standard C++ attributes to introduce annotations tagging the notable components of stream parallel applications: stream sources and stream processing stages. A set of tools process SPar code (C++ annotated code using the SPar attributes) to generate FastFlow C++ code that exploits the stream parallelism denoted by SPar annotations while targeting shared memory multi-core architectures. We outline the main SPar features along with the main implementation techniques and tools. Also, we show the results of experiments assessing the feasibility of the entire approach as well as SPar's performance and expressiveness
Proceedings of the 15th Conference on Knowledge Organization WissOrg'17 of theGerman Chapter of the International Society for Knowledge Organization (ISKO),30th November - 1st December 2017, Freie Universität Berlin
Wissensorganisation is the name of a series of biennial conferences /
workshops with a long tradition, organized by the German chapter of the
International Society of Knowledge Organization (ISKO). The 15th conference in
this series, held at Freie Universität Berlin, focused on knowledge
organization for the digital humanities. Structuring, and interacting with,
large data collections has become a major issue in the digital humanities. In
these proceedings, various aspects of knowledge organization in the digital
humanities are discussed, and the authors of the papers show how projects in
the digital humanities deal with knowledge organization.Wissensorganisation ist der Name einer Konferenzreihe mit einer langjährigen
Tradition, die von der Deutschen Sektion der International Society of
Knowledge Organization (ISKO) organisiert wird. Die 15. Konferenz dieser
Reihe, die an der Freien Universität Berlin stattfand, hatte ihren Schwerpunkt
im Bereich Wissensorganisation und Digital Humanities. Die Strukturierung von
und die Interaktion mit großen Datenmengen ist ein zentrales Thema in den
Digital Humanities. In diesem Konferenzband werden verschiedene Aspekte der
Wissensorganisation in den Digital Humanities diskutiert, und die Autoren der
einzelnen Beiträge zeigen, wie die Digital Humanities mit Wissensorganisation
umgehen
Combining Unsupervised, Supervised, and Rule-based Algorithms for Text Mining of Electronic Health Records - A Clinical Decision Support System for Identifying and Classifying Allergies of Concern for Anesthesia During Surgery
Undisclosed allergic reactions of patients are a major risk when undertaking surgeries in hospitals. We present our early experience and preliminary findings for a Clinical Decision Support System (CDSS) being developed in a Norwegian Hospital Trust. The system incorporates unsupervised and supervised machine learning algorithms in combination with rule-based algorithms to identify and classify allergies of concern for anesthesia during surgery. Our approach is novel in that it utilizes unsupervised machine learning to analyze large corpora of narratives to automatically build a clinical language model containing words and phrases of which meanings and relative meanings are also learnt. It further implements a semi-automatic annotation scheme for efficient and interactive machine-learning, which to a large extent eliminates the substantial manual annotation (of clinical narratives) effort necessary for the training of supervised algorithms. Validation of system performance was performed through comparing allergies identified by the CDSS with a manual reference standard
Prototypicality effects in global semantic description of objects
In this paper, we introduce a novel approach for semantic description of
object features based on the prototypicality effects of the Prototype Theory.
Our prototype-based description model encodes and stores the semantic meaning
of an object, while describing its features using the semantic prototype
computed by CNN-classifications models. Our method uses semantic prototypes to
create discriminative descriptor signatures that describe an object
highlighting its most distinctive features within the category. Our experiments
show that: i) our descriptor preserves the semantic information used by the
CNN-models in classification tasks; ii) our distance metric can be used as the
object's typicality score; iii) our descriptor signatures are semantically
interpretable and enables the simulation of the prototypical organization of
objects within a category.Comment: Paper accepted in IEEE Winter Conference on Applications of Computer
Vision 2019 (WACV2019). Content: 10 pages (8 + 2 reference) with 7 figure
Towards robust and reliable multimedia analysis through semantic integration of services
Thanks to ubiquitous Web connectivity and portable multimedia devices, it has never been so easy to produce and distribute new multimedia resources such as videos, photos, and audio. This ever-increasing production leads to an information overload for consumers, which calls for efficient multimedia retrieval techniques. Multimedia resources can be efficiently retrieved using their metadata, but the multimedia analysis methods that can automatically generate this metadata are currently not reliable enough for highly diverse multimedia content. A reliable and automatic method for analyzing general multimedia content is needed. We introduce a domain-agnostic framework that annotates multimedia resources using currently available multimedia analysis methods. By using a three-step reasoning cycle, this framework can assess and improve the quality of multimedia analysis results, by consecutively (1) combining analysis results effectively, (2) predicting which results might need improvement, and (3) invoking compatible analysis methods to retrieve new results. By using semantic descriptions for the Web services that wrap the multimedia analysis methods, compatible services can be automatically selected. By using additional semantic reasoning on these semantic descriptions, the different services can be repurposed across different use cases. We evaluated this problem-agnostic framework in the context of video face detection, and showed that it is capable of providing the best analysis results regardless of the input video. The proposed methodology can serve as a basis to build a generic multimedia annotation platform, which returns reliable results for diverse multimedia analysis problems. This allows for better metadata generation, and improves the efficient retrieval of multimedia resources
- …