2,126 research outputs found
Requirements for Information Extraction for Knowledge Management
Knowledge Management (KM) systems inherently suffer from the knowledge acquisition bottleneck - the difficulty of modeling and formalizing knowledge relevant for specific domains. A potential solution to this problem is Information Extraction (IE) technology. However, IE was originally developed for database population and there is a mismatch between what is required to successfully perform KM and what current IE technology provides. In this paper we begin to address this issue by outlining requirements for IE based KM
Web based knowledge extraction and consolidation for automatic ontology instantiation
The Web is probably the largest and richest information repository available today. Search engines are the common access routes to this valuable source. However, the role of these search engines is often limited to the retrieval of lists of potentially relevant documents. The burden of analysing the returned documents and identifying the knowledge of interest is therefore left to the user. The Artequakt system aims to deploy natural language tools to automatically ex-tract and consolidate knowledge from web documents and instantiate a given ontology, which dictates the type and form of knowledge to extract. Artequakt focuses on the domain of artists, and uses the harvested knowledge to gen-erate tailored biographies. This paper describes the latest developments of the system and discusses the problem of knowledge consolidation
Recommended from our members
Ontology learning for semantic web services
Semantic Web Services promise automatic service discovery and composition, relying heavily on domain ontology as a core component. With large Web Service repository, manual ontology development is proving a bottleneck (with associated expense and likely errors) to the realisation of a semantic Web of services. Providing the appropriate tools that assist in and automate ontology development is essential for a dynamic service vision to be realised. As a statement of research-in-progress, this paper proposes combining different ontology learning paradigms in Web Services domain, highlighting the need for further research that accommodates the variation in Web Service descriptive and operational sources. A research agenda is proposed that recognises this variation in artefacts as they are selected, pre-processed and analyzed by ontology learning techniques
Language technologies and the evolution of the semantic web
The availability of huge amounts of semantic markup on the Web promises to enable a quantum leap in the level of support available to Web users for locating, aggregating, sharing, interpreting and customizing information. While we cannot claim that a large scale Semantic Web already exists, a number of applications have been produced, which generate and exploit semantic markup, to provide advanced search and querying functionalities, and to allow the visualization and management of heterogeneous, distributed data. While these tools provide evidence of the feasibility and tremendous potential value of the enterprise, they all suffer from major limitations, to do primarily with the limited degree of scale and heterogeneity of the semantic data they use. Nevertheless, we argue that we are at a key point in the brief history of the Semantic Web and that the very latest demonstrators already give us a glimpse of what future applications will look like. In this paper, we describe the already visible effects of these changes by analyzing the evolution of Semantic Web tools from smart databases towards applications that harness collective intelligence. We also point out that language technology plays an important role in making this evolution sustainable and we highlight the need for improved support, especially in the area of large-scale linguistic resources
Unlocking the potential of public sector information with Semantic Web technology
Governments often hold very rich data and whilst much of this information is published and available for re-use by others, it is often trapped by poor data structures, locked up in legacy data formats or in fragmented databases. One of the great benefits that Semantic Web (SW) technology offers is facilitating the large scale integration and sharing of distributed data sources. At the heart of information policy in the UK, the Office of Public Sector Information (OPSI) is the part of the UK government charged with enabling the greater re-use of public sector information. This paper describes the actions, findings, and lessons learnt from a pilot study, involving several parts of government and the public sector. The aim was to show to government how they can adopt SW technology for the dissemination, sharing and use of its data
Automatic annotation of bioinformatics workflows with biomedical ontologies
Legacy scientific workflows, and the services within them, often present
scarce and unstructured (i.e. textual) descriptions. This makes it difficult to
find, share and reuse them, thus dramatically reducing their value to the
community. This paper presents an approach to annotating workflows and their
subcomponents with ontology terms, in an attempt to describe these artifacts in
a structured way. Despite a dearth of even textual descriptions, we
automatically annotated 530 myExperiment bioinformatics-related workflows,
including more than 2600 workflow-associated services, with relevant
ontological terms. Quantitative evaluation of the Information Content of these
terms suggests that, in cases where annotation was possible at all, the
annotation quality was comparable to manually curated bioinformatics resources.Comment: 6th International Symposium on Leveraging Applications (ISoLA 2014
conference), 15 pages, 4 figure
The CKC Challenge: Exploring Tools for Collaborative Knowledge Construction
The great success of Web 2.0 is mainly fuelled by an infrastructure that allows web users to create, share, tag, and connect content and knowledge easily. The tools for developing structured knowledge in this manner have started to appear as well. However, there are few, if any, user studies that are aimed at understanding what users expect from such tools, what works and what doesn't. We organized the Collaborative Knowledge Construction (CKC) Challenge to assess the state of the art for the tools that support collaborative processes for creation of various forms of structured knowledge. The goal of the Challenge was to get users to try out different tools and to learn what users expect from such tools /features that users need, features that they like or dislike. The Challenge task was to construct structured knowledge for a portal that would provide information about research. The Challenge design contained several incentives for users to participate. Forty-nine users registered for the Challenge; thirty three of them participated actively by using the tools. We collected extensive feedback from the users where they discussed their thoughts on all the tools that they tried. In this paper, we present the results of the Challenge, discuss the features that users expect from tools for collaborative knowledge constructions, the features on which Challenge participants disagreed, and the lessons that we learned
TRULLO - local trust bootstrapping for ubiquitous devices
Handheld devices have become sufficiently powerful
that it is easy to create, disseminate, and access digital content
(e.g., photos, videos) using them. The volume of such content is
growing rapidly and, from the perspective of each user, selecting
relevant content is key. To this end, each user may run a trust
model - a software agent that keeps track of who disseminates
content that its user finds relevant. This agent does so by
assigning an initial trust value to each producer for a specific
category (context); then, whenever it receives new content, the
agent rates the content and accordingly updates its trust value for
the producer in the content category. However, a problem with
such an approach is that, as the number of content categories
increases, so does the number of trust values to be initially set.
This paper focuses on how to effectively set initial trust values.
The most sophisticated of the current solutions employ predefined
context ontologies, using which initial trust in a given
context is set based on that already held in similar contexts.
However, universally accepted (and time invariant) ontologies
are rarely found in practice. For this reason, we propose a
mechanism called TRULLO (TRUst bootstrapping by Latently
Lifting cOntext) that assigns initial trust values based only on
local information (on the ratings of its user’s past experiences)
and that, as such, does not rely on third-party recommendations.
We evaluate the effectiveness of TRULLO by simulating its use
in an informal antique market setting. We also evaluate the
computational cost of a J2ME implementation of TRULLO on
a mobile phone
- …