93 research outputs found

    Linked Data - the story so far

    No full text
    The term “Linked Data” refers to a set of best practices for publishing and connecting structured data on the Web. These best practices have been adopted by an increasing number of data providers over the last three years, leading to the creation of a global data space containing billions of assertions— the Web of Data. In this article, the authors present the concept and technical principles of Linked Data, and situate these within the broader context of related technological developments. They describe progress to date in publishing Linked Data on the Web, review applications that have been developed to exploit the Web of Data, and map out a research agenda for the Linked Data community as it moves forward

    X-ENS: Semantic Enrichment of Web Search Results at Real-Time

    Get PDF
    While more and more semantic data are published on the Web, an important question is how typical web users can access and exploit this body of knowledge. Although, existing interaction paradigms in semantic search hide the complexity behind an easy-to-use interface, they have not managed to cover common search needs. In this paper, we present X-ENS (eXplore ENtities in Search), a web search application that enhances the classical, keyword-based, web searching with semantic information, as a means to combine the pros of both Semantic Web standards and common Web Searching. X-ENS identifies entities of interest in the snippets of the top search results which can be further exploited in a faceted interaction scheme, and thereby can help the user to limit the - often very large - search space to those hits that contain a particular piece of information. Moreover, X-ENS permits the exploration of the identified entities by exploiting semantic repositories

    Best Practices of Consuming Linked Open Data

    Get PDF
    The term Linked Data is defined as a set of best practices for publishing and interlinking structured data on the web. These best practices were introduced by Tim Berners-Lee and are also known as principles. These best practices are used by the vast majority of data providers leading to the establishment of a global data space known as the web of data. In this paper will analyze and explore the technical principles of Linked Data, the best practices of using Linked Data, some deployed Linked Data applications and use cases to exploit the Web of Data

    User interfaces supporting entity search for linked data

    Get PDF
    One of the main goals of semantic search is to retrieve and connect information related to queries, offering users rich structured information about a topic instead of a set of documents relevant to the topic. Previous work reports that searching for information about individual entities such as persons, places and organisations is the most common form of Web search. Since the Semantic Web was first proposed, the amount of structured data on the Web has increased dramatically. This is particularly the case for what is known as Linked Data, information that has been published using Semantic Web standards such as RDF and OWL. Such structured data opens up new possibilities for improving entity search on the Web, integrating facts from independent sources, and presenting users with contextually-rich information about entities. This research focuses on entity search of Linked Data in terms of three different forms of search: structured queries, where users can use the SPARQL query language for manipulating data sources; exploratory search, where users can browse from one entity to another; and focused search, where users can input an entity query as a free text keyword search. We undertake a comparative study between two distinct information architectures for structured querying to manipulate Linked Data over the Web. Specifically, we evaluate some of the main operators in SPARQL using several datasets of Linked Data. We introduce a framework of five criteria to evaluate 15 current state-of-the-art semantic tools available for exploratory search of Linked Data, in order to establish how well these browsers make available the benefits of Linked Data and entity search for human users. We also use the criteria to determine the browsers that are best suited to entity exploration. Further, we propose a new model, the Attribute Importance Model, for entity-aggregated search, with the purpose of improving user experience when finding information about entities. The model develops three techniques: (1) presenting entity type-based query suggestions; (2) clustering aggregated attributes; and (3) ranking attributes based on their importance to a given query. Together these constitute a model for developing more informative views and enhancing users’ understanding of entity descriptions on the Web. We then use our model to provide an interactive approach, with the Information Visualisation toolkit InfoVis, that enables users to visualise entity clusters generated by our Attribute Importance Model. Thus this thesis addresses two challenges of searching Linked Data. The first challenge concerns the specific issue of information resolution during the search: the reduction of query ambiguity and redundant results that contain irrelevant descriptions when searching for information about an entity. The second challenge concerns the more general problem of technical complexity, and addresses to the limited adoption of Linked Data that we ascribe to the lack of understanding of Semantic Web technologies and data structures among general users. These technologies pose new design problems for human interaction such as overloading data, navigation styles, and browsing mechanisms. The Attribute Importance Model addresses both these challenges

    Representing Dataset Quality Metadata using Multi-Dimensional Views

    Full text link
    Data quality is commonly defined as fitness for use. The problem of identifying quality of data is faced by many data consumers. Data publishers often do not have the means to identify quality problems in their data. To make the task for both stakeholders easier, we have developed the Dataset Quality Ontology (daQ). daQ is a core vocabulary for representing the results of quality benchmarking of a linked dataset. It represents quality metadata as multi-dimensional and statistical observations using the Data Cube vocabulary. Quality metadata are organised as a self-contained graph, which can, e.g., be embedded into linked open datasets. We discuss the design considerations, give examples for extending daQ by custom quality metrics, and present use cases such as analysing data versions, browsing datasets by quality, and link identification. We finally discuss how data cube visualisation tools enable data publishers and consumers to analyse better the quality of their data.Comment: Preprint of a paper submitted to the forthcoming SEMANTiCS 2014, 4-5 September 2014, Leipzig, German

    DBpedia Mashups

    Get PDF
    If you see Wikipedia as a main place where the knowledge of mankind is concentrated, then DBpedia – which is extracted from Wikipedia – is the best place to find machine representation of that knowledge. DBpedia constitutes a major part of the semantic data on the web. Its sheer size and wide coverage enables you to use it in many kind of mashups: it contains biographical, geographical, bibliographical data; as well as discographies, movie meta-data, technical specifications, and links to social media profiles and much more. Just like Wikipedia, DBpedia is a truly cross-language effort, e.g., it provides descriptions and other information in various languages. In this chapter we introduce its structure, contents, its connections to outside resources. We describe how the structured information in DBpedia is gathered, what you can expect from it and what are its characteristics and limitations. We analyze how other mashups exploit DBpedia and present best practices of its usage. In particular, we describe how Sztakipedia – an intelligent writing aid based on DBpedia – can help Wikipedia contributors to improve the quality and integrity of articles. DBpedia offers a myriad of ways to accessing the information it contains, ranging from SPARQL to bulk download. We compare the pros and cons of these methods. We conclude that DBpedia is an un-avoidable resource for pplications dealing with commonly known entities like notable persons, places; and for others looking for a rich hub connecting other semantic resources
    • …
    corecore