13,762 research outputs found
Investigations on Knowledge Base Embedding for Relation Prediction and Extraction
We report an evaluation of the effectiveness of the existing knowledge base
embedding models for relation prediction and for relation extraction on a wide
range of benchmarks. We also describe a new benchmark, which is much larger and
complex than previous ones, which we introduce to help validate the
effectiveness of both tasks. The results demonstrate that knowledge base
embedding models are generally effective for relation prediction but unable to
give improvements for the state-of-art neural relation extraction model with
the existing strategies, while pointing limitations of existing methods
Improving Visual Relation Detection using Depth Maps
State-of-the-art visual relation detection methods mostly rely on object
information extracted from RGB images such as predicted class probabilities, 2D
bounding boxes and feature maps. Depth maps can additionally provide valuable
information on object relations, e.g. helping to detect not only spatial
relations, such as standing behind, but also non-spatial relations, such as
holding. In this work, we study the effect of using different object
information with a focus on depth maps. To enable this study, we release a new
synthetic dataset of depth maps, VG-Depth, as an extension to Visual Genome
(VG). We also note that given the highly imbalanced distribution of relations
in VG, typical evaluation metrics for visual relation detection cannot reveal
improvements of under-represented relations. To address this problem, we
propose using an additional metric, calling it Macro Recall@K, and demonstrate
its remarkable performance on VG. Finally, our experiments confirm that by
effective utilization of depth maps within a simple, yet competitive framework,
the performance of visual relation detection can be significantly improved
Indirect Supervision for Relation Extraction using Question-Answer Pairs
Automatic relation extraction (RE) for types of interest is of great
importance for interpreting massive text corpora in an efficient manner.
Traditional RE models have heavily relied on human-annotated corpus for
training, which can be costly in generating labeled data and become obstacles
when dealing with more relation types. Thus, more RE extraction systems have
shifted to be built upon training data automatically acquired by linking to
knowledge bases (distant supervision). However, due to the incompleteness of
knowledge bases and the context-agnostic labeling, the training data collected
via distant supervision (DS) can be very noisy. In recent years, as increasing
attention has been brought to tackling question-answering (QA) tasks, user
feedback or datasets of such tasks become more accessible. In this paper, we
propose a novel framework, ReQuest, to leverage question-answer pairs as an
indirect source of supervision for relation extraction, and study how to use
such supervision to reduce noise induced from DS. Our model jointly embeds
relation mentions, types, QA entity mention pairs and text features in two
low-dimensional spaces (RE and QA), where objects with same relation types or
semantically similar question-answer pairs have similar representations. Shared
features connect these two spaces, carrying clearer semantic knowledge from
both sources. ReQuest, then use these learned embeddings to estimate the types
of test relation mentions. We formulate a global objective function and adopt a
novel margin-based QA loss to reduce noise in DS by exploiting semantic
evidence from the QA dataset. Our experimental results achieve an average of
11% improvement in F1 score on two public RE datasets combined with TREC QA
dataset.Comment: 9 pages + 1 page reference. Accepted to WSDM 201
EMR-based medical knowledge representation and inference via Markov random fields and distributed representation learning
Objective: Electronic medical records (EMRs) contain an amount of medical
knowledge which can be used for clinical decision support (CDS). Our objective
is a general system that can extract and represent these knowledge contained in
EMRs to support three CDS tasks: test recommendation, initial diagnosis, and
treatment plan recommendation, with the given condition of one patient.
Methods: We extracted four kinds of medical entities from records and
constructed an EMR-based medical knowledge network (EMKN), in which nodes are
entities and edges reflect their co-occurrence in a single record. Three
bipartite subgraphs (bi-graphs) were extracted from the EMKN to support each
task. One part of the bi-graph was the given condition (e.g., symptoms), and
the other was the condition to be inferred (e.g., diseases). Each bi-graph was
regarded as a Markov random field to support the inference. Three lazy energy
functions and one parameter-based energy function were proposed, as well as two
knowledge representation learning-based energy functions, which can provide a
distributed representation of medical entities. Three measures were utilized
for performance evaluation. Results: On the initial diagnosis task, 80.11% of
the test records identified at least one correct disease from top 10
candidates. Test and treatment recommendation results were 87.88% and 92.55%,
respectively. These results altogether indicate that the proposed system
outperformed the baseline methods. The distributed representation of medical
entities does reflect similarity relationships in regards to knowledge level.
Conclusion: Combining EMKN and MRF is an effective approach for general medical
knowledge representation and inference. Different tasks, however, require
designing their energy functions individually
Noise-robust Named Entity Understanding for Virtual Assistants
Named Entity Understanding (NEU) plays an essential role in interactions
between users and voice assistants, since successfully identifying entities and
correctly linking them to their standard forms is crucial to understanding the
user's intent. NEU is a challenging task in voice assistants due to the
ambiguous nature of natural language and because noise introduced by speech
transcription and user errors occur frequently in spoken natural language
queries. In this paper, we propose an architecture with novel features that
jointly solves the recognition of named entities (a.k.a. Named Entity
Recognition, or NER) and the resolution to their canonical forms (a.k.a. Entity
Linking, or EL). We show that by combining NER and EL information in a joint
reranking module, our proposed framework improves accuracy in both tasks. This
improved performance and the features that enable it, also lead to better
accuracy in downstream tasks, such as domain classification and semantic
parsing.Comment: 9 page
Joint Matrix-Tensor Factorization for Knowledge Base Inference
While several matrix factorization (MF) and tensor factorization (TF) models
have been proposed for knowledge base (KB) inference, they have rarely been
compared across various datasets. Is there a single model that performs well
across datasets? If not, what characteristics of a dataset determine the
performance of MF and TF models? Is there a joint TF+MF model that performs
robustly on all datasets? We perform an extensive evaluation to compare popular
KB inference models across popular datasets in the literature. In addition to
answering the questions above, we remove a limitation in the standard
evaluation protocol for MF models, propose an extension to MF models so that
they can better handle out-of-vocabulary (OOV) entity pairs, and develop a
novel combination of TF and MF models. We also analyze and explain the results
based on models and dataset characteristics. Our best model is robust, and
obtains strong results across all datasets
Deep Learning applied to NLP
Convolutional Neural Network (CNNs) are typically associated with Computer
Vision. CNNs are responsible for major breakthroughs in Image Classification
and are the core of most Computer Vision systems today. More recently CNNs have
been applied to problems in Natural Language Processing and gotten some
interesting results. In this paper, we will try to explain the basics of CNNs,
its different variations and how they have been applied to NLP
Encoding Implicit Relation Requirements for Relation Extraction: A Joint Inference Approach
Relation extraction is the task of identifying predefined relationship
between entities, and plays an essential role in information extraction,
knowledge base construction, question answering and so on. Most existing
relation extractors make predictions for each entity pair locally and
individually, while ignoring implicit global clues available across different
entity pairs and in the knowledge base, which often leads to conflicts among
local predictions from different entity pairs. This paper proposes a joint
inference framework that employs such global clues to resolve disagreements
among local predictions. We exploit two kinds of clues to generate constraints
which can capture the implicit type and cardinality requirements of a relation.
Those constraints can be examined in either hard style or soft style, both of
which can be effectively explored in an integer linear program formulation.
Experimental results on both English and Chinese datasets show that our
proposed framework can effectively utilize those two categories of global clues
and resolve the disagreements among local predictions, thus improve various
relation extractors when such clues are applicable to the datasets. Our
experiments also indicate that the clues learnt automatically from existing
knowledge bases perform comparably to or better than those refined by human.Comment: to appear in Artificial Intelligenc
Reasoning over RDF Knowledge Bases using Deep Learning
Semantic Web knowledge representation standards, and in particular RDF and
OWL, often come endowed with a formal semantics which is considered to be of
fundamental importance for the field. Reasoning, i.e., the drawing of logical
inferences from knowledge expressed in such standards, is traditionally based
on logical deductive methods and algorithms which can be proven to be sound and
complete and terminating, i.e. correct in a very strong sense. For various
reasons, though, in particular, the scalability issues arising from the
ever-increasing amounts of Semantic Web data available and the inability of
deductive algorithms to deal with noise in the data, it has been argued that
alternative means of reasoning should be investigated which bear high promise
for high scalability and better robustness. From this perspective, deductive
algorithms can be considered the gold standard regarding correctness against
which alternative methods need to be tested. In this paper, we show that it is
possible to train a Deep Learning system on RDF knowledge graphs, such that it
is able to perform reasoning over new RDF knowledge graphs, with high precision
and recall compared to the deductive gold standard
Take and Took, Gaggle and Goose, Book and Read: Evaluating the Utility of Vector Differences for Lexical Relation Learning
Recent work on word embeddings has shown that simple vector subtraction over
pre-trained embeddings is surprisingly effective at capturing different lexical
relations, despite lacking explicit supervision. Prior work has evaluated this
intriguing result using a word analogy prediction formulation and hand-selected
relations, but the generality of the finding over a broader range of lexical
relation types and different learning settings has not been evaluated. In this
paper, we carry out such an evaluation in two learning settings: (1) spectral
clustering to induce word relations, and (2) supervised learning to classify
vector differences into relation types. We find that word embeddings capture a
surprising amount of information, and that, under suitable supervised training,
vector subtraction generalises well to a broad range of relations, including
over unseen lexical items
- …