21,561 research outputs found
Deductive and Analogical Reasoning on a Semantically Embedded Knowledge Graph
Representing knowledge as high-dimensional vectors in a continuous semantic
vector space can help overcome the brittleness and incompleteness of
traditional knowledge bases. We present a method for performing deductive
reasoning directly in such a vector space, combining analogy, association, and
deduction in a straightforward way at each step in a chain of reasoning,
drawing on knowledge from diverse sources and ontologies.Comment: AGI 201
Interpreting Embedding Models of Knowledge Bases: A Pedagogical Approach
Knowledge bases are employed in a variety of applications from natural
language processing to semantic web search; alas, in practice their usefulness
is hurt by their incompleteness. Embedding models attain state-of-the-art
accuracy in knowledge base completion, but their predictions are notoriously
hard to interpret. In this paper, we adapt "pedagogical approaches" (from the
literature on neural networks) so as to interpret embedding models by
extracting weighted Horn rules from them. We show how pedagogical approaches
have to be adapted to take upon the large-scale relational aspects of knowledge
bases and show experimentally their strengths and weaknesses.Comment: presented at 2018 ICML Workshop on Human Interpretability in Machine
Learning (WHI 2018), Stockholm, Swede
Empowering Knowledge Bases: a Machine Learning Perspective
The construction of Knowledge Bases requires quite often
the intervention of knowledge engineering and domain experts, resulting
in a time consuming task. Alternative approaches have been developed
for building knowledge bases from existing sources of information such
as web pages and crowdsourcing; seminal examples are NELL, DBPedia,
YAGO and several others. With the goal of building very large sources of
knowledge, as recently for the case of Knowledge Graphs, even more complex
integration processes have been set up, involving multiple sources of
information, human expert intervention, crowdsourcing. Despite signi -
cant e orts for making Knowledge Graphs as comprehensive and reliable
as possible, they tend to su er of incompleteness and noise, due to the
complex building process. Nevertheless, even for highly human curated
knowledge bases, cases of incompleteness can be found, for instance with
disjointness axioms missing quite often. Machine learning methods have
been proposed with the purpose of re ning, enriching, completing and
possibly raising potential issues in existing knowledge bases while showing
the ability to cope with noise. The talk will concentrate on classes
of mostly symbol-based machine learning methods, speci cally focusing
on concept learning, rule learning and disjointness axioms learning problems,
showing how the developed methods can be exploited for enriching
existing knowledge bases. During the talk it will be highlighted as, a
key element of the illustrated solutions, is represented by the integration
of: background knowledge, deductive reasoning and the evidence coming
from the mass of the data. The last part of the talk will be devoted
to the presentation of an approach for injecting background knowledge
into numeric-based embedding models to be used for predictive tasks on
Knowledge Graphs
Utilizing Data and Knowledge Mining for Probabilistic Knowledge Bases
Problems can arise whenever inferencing is attempted on a knowledge base that is incomplete. Our work shows that data mining techniques can be applied to fill in incomplete areas in Bayesian Knowledge Bases (BKBs), as well as in other knowledge-based systems utilizing probabilistic representations. The problem of inconsistency in BKBs has been addressed in previous work, where reinforcement learning techniques from neural networks were applied. However, the issue of automatically solving incompleteness in BKBs has yet to be addressed. Presently, incompleteness in BKBs is repaired through the application of traditional knowledge acquisition techniques. We show how association rules can be extracted from databases in order to replace excluded information and express missing relationships. A methodology for incorporating those results while maintaining a consistent knowledge base is also included
Learning New Facts From Knowledge Bases With Neural Tensor Networks and Semantic Word Vectors
Knowledge bases provide applications with the benefit of easily accessible,
systematic relational knowledge but often suffer in practice from their
incompleteness and lack of knowledge of new entities and relations. Much work
has focused on building or extending them by finding patterns in large
unannotated text corpora. In contrast, here we mainly aim to complete a
knowledge base by predicting additional true relationships between entities,
based on generalizations that can be discerned in the given knowledgebase. We
introduce a neural tensor network (NTN) model which predicts new relationship
entries that can be added to the database. This model can be improved by
initializing entity representations with word vectors learned in an
unsupervised fashion from text, and when doing this, existing relations can
even be queried for entities that were not present in the database. Our model
generalizes and outperforms existing models for this problem, and can classify
unseen relationships in WordNet with an accuracy of 75.8%
Approximate declarative semantics for rule base anomalies
The article of record as published may be found at https://doi.org/10.1016/S0950-7051(99)00032-5Despite the fact that there has been a surge of publications in verification and validation of knowledge-based systems and expert systems in the past decade, there are still gaps in the study of verification and validation (V&V) of expert systems, not the least of which is the lack of appropriate semantics for expert system programming languages. Without a semantics, it is hard to formally define and analyze knowledge base anomalies such as inconsistency and redundancy, and it is hard to assess the effectiveness of V&V tools, methods and techniques that have been developed or proposed. In this paper, we develop an approximate declarative semantics for rule-based knowledge bases and provide a formal definition and analysis of knowledge base inconsistency, redundancy, circularity and incompleteness in terms of theories in the first order predicate logic. In the paper, we offer classifications of commonly found cases of inconsistency, redundancy, circularity and incompleteness. Finally, general guidelines on how to remedy knowledge base anomalies are given
Method for Detecting Anomalous States of a Control Object in Information Systems Based on the Analysis of Temporal Data and Knowledge
The problem of finding the anomalous states of the control object in the management information system under conditions of uncertainty caused by the incompleteness of knowledge about this object is considered. The method of classifying the current state of the control object in real time, allowing to identify the current anomalous state. The method uses temporal data and knowledge. Data is represented by sequences of events with timestamps. Knowledge is represented as weighted temporal rules and constraints. The method includes the following key phases: the formation of sequences of logical facts; selection of temporal rules and constraints; classification based on a comparison of rules and constraints. Logical facts are represented as predicates on event attributes and reflect the state of the control object. Logical rules define valid sequences of logical facts. Performing a classification by successive comparisons of constraints and weights of the rules makes it possible to more effectively identify the anomalous state since the comparison of the constraints reduces the subset of facts comparing to the current state. The method creates conditions for improving management efficiency in the context of incomplete information on the state of a complex object by using logical inference in knowledge bases for anomalous states of such control objects
Enriching Knowledge Bases with Counting Quantifiers
Information extraction traditionally focuses on extracting relations between
identifiable entities, such as . Yet, texts
often also contain Counting information, stating that a subject is in a
specific relation with a number of objects, without mentioning the objects
themselves, for example, "California is divided into 58 counties". Such
counting quantifiers can help in a variety of tasks such as query answering or
knowledge base curation, but are neglected by prior work. This paper develops
the first full-fledged system for extracting counting information from text,
called CINEX. We employ distant supervision using fact counts from a knowledge
base as training seeds, and develop novel techniques for dealing with several
challenges: (i) non-maximal training seeds due to the incompleteness of
knowledge bases, (ii) sparse and skewed observations in text sources, and (iii)
high diversity of linguistic patterns. Experiments with five human-evaluated
relations show that CINEX can achieve 60% average precision for extracting
counting information. In a large-scale experiment, we demonstrate the potential
for knowledge base enrichment by applying CINEX to 2,474 frequent relations in
Wikidata. CINEX can assert the existence of 2.5M facts for 110 distinct
relations, which is 28% more than the existing Wikidata facts for these
relations.Comment: 16 pages, The 17th International Semantic Web Conference (ISWC 2018
- …