23 research outputs found
Context-aware Human Motion Prediction
The problem of predicting human motion given a sequence of past observations
is at the core of many applications in robotics and computer vision. Current
state-of-the-art formulate this problem as a sequence-to-sequence task, in
which a historical of 3D skeletons feeds a Recurrent Neural Network (RNN) that
predicts future movements, typically in the order of 1 to 2 seconds. However,
one aspect that has been obviated so far, is the fact that human motion is
inherently driven by interactions with objects and/or other humans in the
environment. In this paper, we explore this scenario using a novel
context-aware motion prediction architecture. We use a semantic-graph model
where the nodes parameterize the human and objects in the scene and the edges
their mutual interactions. These interactions are iteratively learned through a
graph attention layer, fed with the past observations, which now include both
object and human body motions. Once this semantic graph is learned, we inject
it to a standard RNN to predict future movements of the human/s and object/s.
We consider two variants of our architecture, either freezing the contextual
interactions in the future of updating them. A thorough evaluation in the
"Whole-Body Human Motion Database" shows that in both cases, our context-aware
networks clearly outperform baselines in which the context information is not
considered.Comment: Accepted at CVPR2
Policy based roles for distributed systems security
Distributed systems are increasingly being used in commercial environments necessitating the development of trustworthy and reliable security mechanisms. There is often no clear informal or formal specification of enterprise authorisation policies and no tools to translate policy specifications to access control implementation mechanisms such as capabilities or Access Control Lists. It is thus difficult to analyse the policy to detect conflicts or flaws and it is difficult to verify that the implementation corresponds to the policy specification. We present in this paper a framework for the specification of management policies. We are concerned with two types of policies: obligations which specify what activities a manager or agent must or must not perform on a set of target objects and authorisations which specify what activities a subject (manager or agent) can or can not perform on the set of target objects. Management policies are then grouped into roles reflecting the organisation..
Semantic Search Engine as tool for clinical decision support in Register for Acute Coronary Syndrome
This paper presents the implementation and use of Semantic Search Engine (SSE) as part of knowledge management system functionalities in Register for Acute Coronary Syndrome (REACS). REACS SSE is part of a clinical decision support system and is used as an aid in decision making in clinical processes related to the care and treatment of patients with Acute Coronary Syndrome (ACS)
Graph based text representation for document clustering
Advances in digital technology and the World Wide Web has led to the increase of digital documents that are used for various purposes such as publishing and digital library. This phenomenon raises awareness for the requirement of effective techniques that can help during the search and retrieval of text. One of the most needed tasks is clustering, which categorizes documents automatically into
meaningful groups. Clustering is an important task in data mining and machine learning. The accuracy of clustering depends tightly on the selection of the text representation method. Traditional methods of text representation model documents as bags of words using term-frequency index document frequency (TFIDF). This method ignores the relationship and meanings of words in the document. As a result the sparsity and semantic problem that is prevalent in textual document are not
resolved. In this study, the problem of sparsity and semantic is reduced by proposing a graph based text representation method, namely dependency graph with the aim of improving the accuracy of document clustering. The dependency graph representation scheme is created through an accumulation of syntactic and semantic
analysis. A sample of 20 news group, dataset was used in this study. The text documents undergo pre-processing and syntactic parsing in order to identify the sentence structure. Then the semantic of words are modeled using dependency graph. The produced dependency graph is then used in the process of cluster analysis. K-means clustering technique was used in this study. The dependency graph based clustering result were compared with the popular text representation method, i.e. TFIDF and Ontology based text representation. The result shows that the dependency graph outperforms both TFIDF and Ontology based text
representation. The findings proved that the proposed text representation method leads to more accurate document clustering results
LinkHub: a Semantic Web system that facilitates cross-database queries and information retrieval in proteomics
<p>Abstract</p> <p>Background</p> <p>A key abstraction in representing proteomics knowledge is the notion of unique identifiers for individual entities (e.g. proteins) and the massive graph of relationships among them. These relationships are sometimes simple (e.g. synonyms) but are often more complex (e.g. one-to-many relationships in protein family membership).</p> <p>Results</p> <p>We have built a software system called LinkHub using Semantic Web RDF that manages the graph of identifier relationships and allows exploration with a variety of interfaces. For efficiency, we also provide relational-database access and translation between the relational and RDF versions. LinkHub is practically useful in creating small, local hubs on common topics and then connecting these to major portals in a federated architecture; we have used LinkHub to establish such a relationship between UniProt and the North East Structural Genomics Consortium. LinkHub also facilitates queries and access to information and documents related to identifiers spread across multiple databases, acting as "connecting glue" between different identifier spaces. We demonstrate this with example queries discovering "interologs" of yeast protein interactions in the worm and exploring the relationship between gene essentiality and pseudogene content. We also show how "protein family based" retrieval of documents can be achieved. LinkHub is available at hub.gersteinlab.org and hub.nesg.org with supplement, database models and full-source code.</p> <p>Conclusion</p> <p>LinkHub leverages Semantic Web standards-based integrated data to provide novel information retrieval to identifier-related documents through relational graph queries, simplifies and manages connections to major hubs such as UniProt, and provides useful interactive and query interfaces for exploring the integrated data.</p
Contact-aware Human Motion Forecasting
In this paper, we tackle the task of scene-aware 3D human motion forecasting,
which consists of predicting future human poses given a 3D scene and a past
human motion. A key challenge of this task is to ensure consistency between the
human and the scene, accounting for human-scene interactions. Previous attempts
to do so model such interactions only implicitly, and thus tend to produce
artifacts such as "ghost motion" because of the lack of explicit constraints
between the local poses and the global motion. Here, by contrast, we propose to
explicitly model the human-scene contacts. To this end, we introduce
distance-based contact maps that capture the contact relationships between
every joint and every 3D scene point at each time instant. We then develop a
two-stage pipeline that first predicts the future contact maps from the past
ones and the scene point cloud, and then forecasts the future human poses by
conditioning them on the predicted contact maps. During training, we explicitly
encourage consistency between the global motion and the local poses via a prior
defined using the contact maps and future poses. Our approach outperforms the
state-of-the-art human motion forecasting and human synthesis methods on both
synthetic and real datasets. Our code is available at
https://github.com/wei-mao-2019/ContAwareMotionPred.Comment: Accepted to NeurIPS202
Modular Design Patterns for Hybrid Learning and Reasoning Systems: a taxonomy, patterns and use cases
The unification of statistical (data-driven) and symbolic (knowledge-driven)
methods is widely recognised as one of the key challenges of modern AI. Recent
years have seen large number of publications on such hybrid neuro-symbolic AI
systems. That rapidly growing literature is highly diverse and mostly
empirical, and is lacking a unifying view of the large variety of these hybrid
systems. In this paper we analyse a large body of recent literature and we
propose a set of modular design patterns for such hybrid, neuro-symbolic
systems. We are able to describe the architecture of a very large number of
hybrid systems by composing only a small set of elementary patterns as building
blocks.
The main contributions of this paper are: 1) a taxonomically organised
vocabulary to describe both processes and data structures used in hybrid
systems; 2) a set of 15+ design patterns for hybrid AI systems, organised in a
set of elementary patterns and a set of compositional patterns; 3) an
application of these design patterns in two realistic use-cases for hybrid AI
systems. Our patterns reveal similarities between systems that were not
recognised until now. Finally, our design patterns extend and refine Kautz'
earlier attempt at categorising neuro-symbolic architectures.Comment: 20 pages, 22 figures, accepted for publication in the International
Journal of Applied Intelligenc
Recommended from our members
Semantic units: organizing knowledge graphs into semantically meaningful units of representation
Background
In today’s landscape of data management, the importance of knowledge graphs and ontologies is escalating as critical mechanisms aligned with the FAIR Guiding Principles—ensuring data and metadata are Findable, Accessible, Interoperable, and Reusable. We discuss three challenges that may hinder the effective exploitation of the full potential of FAIR knowledge graphs.
Results
We introduce “semantic units” as a conceptual solution, although currently exemplified only in a limited prototype. Semantic units structure a knowledge graph into identifiable and semantically meaningful subgraphs by adding another layer of triples on top of the conventional data layer. Semantic units and their subgraphs are represented by their own resource that instantiates a corresponding semantic unit class. We distinguish statement and compound units as basic categories of semantic units. A statement unit is the smallest, independent proposition that is semantically meaningful for a human reader. Depending on the relation of its underlying proposition, it consists of one or more triples. Organizing a knowledge graph into statement units results in a partition of the graph, with each triple belonging to exactly one statement unit. A compound unit, on the other hand, is a semantically meaningful collection of statement and compound units that form larger subgraphs. Some semantic units organize the graph into different levels of representational granularity, others orthogonally into different types of granularity trees or different frames of reference, structuring and organizing the knowledge graph into partially overlapping, partially enclosed subgraphs, each of which can be referenced by its own resource.
Conclusions
Semantic units, applicable in RDF/OWL and labeled property graphs, offer support for making statements about statements and facilitate graph-alignment, subgraph-matching, knowledge graph profiling, and for management of access restrictions to sensitive data. Additionally, we argue that organizing the graph into semantic units promotes the differentiation of ontological and discursive information, and that it also supports the differentiation of multiple frames of reference within the graph