24 research outputs found
Towards Knowledge in the Cloud
Knowledge in the form of semantic data is becoming more and more ubiquitous, and the need for scalable, dynamic systems to support collaborative work with such distributed, heterogeneous knowledge arises. We extend the “data in the cloud” approach that is emerging today to “knowledge in the cloud”, with support for handling semantic information, organizing and finding it efficiently and providing reasoning and quality support. Both the life sciences and emergency response fields are identified as strong potential beneficiaries of having ”knowledge in the cloud”
RelBAC: Relation Based Access Control
TheWeb 2.0, GRID applications and, more recently, semantic desktop applications are bringing the Web to a situation where more and more data and metadata are shared and made available to large user groups. In this context, metadata may be tags or complex graph structures such as file system or web directories, or (lightweight) ontologies. In turn, users can themselves be tagged by certain properties, and can be organized in complex directory structures, very much in the same way as data. Things are further complicated by the highly unpredictable and autonomous dynamics of data, users, permissions and access control rules. In this paper we propose a new access control model and a logic, called RelBAC (for Relation Based Access Control) which allows us to deal with this novel scenario. The key idea, which differentiates RelBAC from the state of the art, e.g., Role Based Access Control (RBAC), is that permissions are modeled as relations between users and data, while access control rules are their instantiations on specific sets of users and objects. As such, access control rules are assigned an arity which allows a fine tuning of which users can access which data, and can evolve independently, according to the desires of the policy manager(s). Furthermore, the formalization of the RelBAC model as an Entity-Relationship (ER) model allows for its direct translation into Description Logics (DL). In turn, this allows us to reason, possibly at run time, about access control policies
A Large Scale Dataset for the Evaluation of Ontology Matching Systems
Recently, the number of ontology matching techniques and systems has increased significantly. This makes the issue of their evaluation and comparison more severe. One of the challenges of the ontology matching evaluation is in building large scale evaluation datasets. In fact, the number of possible correspondences between two ontologies grows quadratically with respect to the numbers of entities in these ontologies. This often makes the manual construction of the evaluation datasets demanding to the point of being infeasible for large scale matching tasks. In this paper we present an ontology matching evaluation dataset composed of thousands of matching tasks, called TaxME2. It was built semi-automatically out of the Google, Yahoo and Looksmart web directories. We evaluated TaxME2 by exploiting the results of almost two dozen of state of the art ontology matching systems. The experiments indicate that the dataset possesses the desired key properties, namely it is error-free, incremental, discriminative, monotonic, and hard for the state of the art ontology matching systems. The paper has been accepted for publication in "The Knowledge Engineering Review", Cambridge Universty Press (ISSN: 0269-8889, EISSN: 1469-8005)
An Analysis of Service Trading Architectures
Automating the creation and management of SLAs in elec tronic commerce scenarios brings many advantages, such as increasing
the speed in the contracting process or allowing providers to deploy an
automated provision of services based on those SLAs. We focus on the
service trading process, which is the process of locating, selecting, nego tiating, and creating SLAs. This process can be applied to a variety of
scenarios and, hence, their requirements are also very different. Despite
some service trading architectures have been proposed, currently there is
no analysis about which one fits better in each scenario. In this paper, we
define a set of properties for abstract service trading architectures based
on an analysis of several practical scenarios. Then, we use it to analyse
and compare the most relevant abstract architectures for service trad ing. In so doing, the main contribution of this article is a first approach
to settle the basis for a qualitative selection of the best architecture for
similar trading scenarios
Ontologie leggere a faccette
In questo articolo ci concentriamo sull’uso delle ontologie per l’organizzazione di oggetti, quali ad esempio foto, libri e pagine Web. Le ontologie leggere sono ontologie con una struttura gerarchica ad albero dove a ciascun nodo è associata un’etichetta in linguaggio naturale. Nelle ontologie leggere a faccette le etichette sono organizzate secondo modelli ben definiti, i quali catturano specifici aspetti della conoscenza, ovvero le faccette. A tal fine, ci basiamo sull’approccio Analitico-Sintetico, una ben radicata metodologia usata con successo per decenni in biblioteconomia, soprattutto in India, per la classificazione di libri. Le ontologie leggere a faccette hanno una struttura ben definita ed, in quanto tali, risultano più facili da creare, condividere tra gli utenti, e più appropriate in applicazioni semantiche, dove cioè viene automaticamente analizzato e sfruttato il significato ontologico dei termini
Survey: Models and Prototypes of Schema Matching
Schema matching is critical problem within many applications to integration of data/information, to achieve interoperability, and other cases caused by schematic heterogeneity. Schema matching evolved from manual way on a specific domain, leading to a new models and methods that are semi-automatic and more general, so it is able to effectively direct the user within generate a mapping among elements of two the schema or ontologies better. This paper is a summary of literature review on models and prototypes on schema matching within the last 25 years to describe the progress of and research chalenge and opportunities on a new models, methods, and/or prototypes
Efficient Neural Ranking using Forward Indexes and Lightweight Encoders
Dual-encoder-based dense retrieval models have become the standard in IR.
They employ large Transformer-based language models, which are notoriously
inefficient in terms of resources and latency. We propose Fast-Forward indexes
-- vector forward indexes which exploit the semantic matching capabilities of
dual-encoder models for efficient and effective re-ranking. Our framework
enables re-ranking at very high retrieval depths and combines the merits of
both lexical and semantic matching via score interpolation. Furthermore, in
order to mitigate the limitations of dual-encoders, we tackle two main
challenges: Firstly, we improve computational efficiency by either
pre-computing representations, avoiding unnecessary computations altogether, or
reducing the complexity of encoders. This allows us to considerably improve
ranking efficiency and latency. Secondly, we optimize the memory footprint and
maintenance cost of indexes; we propose two complementary techniques to reduce
the index size and show that, by dynamically dropping irrelevant document
tokens, the index maintenance efficiency can be improved substantially. We
perform evaluation to show the effectiveness and efficiency of Fast-Forward
indexes -- our method has low latency and achieves competitive results without
the need for hardware acceleration, such as GPUs.Comment: Accepted at ACM TOIS. arXiv admin note: text overlap with
arXiv:2110.0605
Discovering Missing Background Knowledge in Ontology Matching
Semantic matching determines the mappings between the nodes of two graphs (e.g., ontologies) by computing logical relations (e.g., subsumption) holding among the nodes that correspond semantically to each other. We present an approach to deal with the lack of background knowledge in matching tasks by using semantic matching iteratively. Unlike previous approaches, where the missing axioms are manually declared before the matching starts, we propose a fully automated solution. The benefits of our approach are: (i) saving some of the pre-match efforts, (ii) improving the quality of match via iterations, and (iii) enabling the future reuse of the newly discovered knowledge. We evaluate the implemented system on large real-world test cases, thus, proving empirically the benefits of our approach