11,723 research outputs found
Hypermedia-based discovery for source selection using low-cost linked data interfaces
Evaluating federated Linked Data queries requires consulting multiple sources on the Web. Before a client can execute queries, it must discover data sources, and determine which ones are relevant. Federated query execution research focuses on the actual execution, while data source discovery is often marginally discussed-even though it has a strong impact on selecting sources that contribute to the query results. Therefore, the authors introduce a discovery approach for Linked Data interfaces based on hypermedia links and controls, and apply it to federated query execution with Triple Pattern Fragments. In addition, the authors identify quantitative metrics to evaluate this discovery approach. This article describes generic evaluation measures and results for their concrete approach. With low-cost data summaries as seed, interfaces to eight large real-world datasets can discover each other within 7 minutes. Hypermedia-based client-side querying shows a promising gain of up to 50% in execution time, but demands algorithms that visit a higher number of interfaces to improve result completeness
Efficient Query Processing for SPARQL Federations with Replicated Fragments
Low reliability and availability of public SPARQL endpoints prevent
real-world applications from exploiting all the potential of these querying
infras-tructures. Fragmenting data on servers can improve data availability but
degrades performance. Replicating fragments can offer new tradeoff between
performance and availability. We propose FEDRA, a framework for querying Linked
Data that takes advantage of client-side data replication, and performs a
source selection algorithm that aims to reduce the number of selected public
SPARQL endpoints, execution time, and intermediate results. FEDRA has been
implemented on the state-of-the-art query engines ANAPSID and FedX, and
empirically evaluated on a variety of real-world datasets
Context Aware Source Selection for Linked Data
The traditional Web is evolving into the Web of Data, which gathers huge collections of structured data over distributed, heterogeneous data sources. Live queries are needed to get current information out of this global data space. In live query processing, source selection allows the identification of the sources that most likely contain relevant content. Due to the semantic heterogeneity of the Web of Data, however, it is not always easy to assess relevancy. Context information might help in interpreting the user\u2019s information needs. In this paper, we discuss how context information can be exploited to improve source selection
Exploiting Context-Dependent Quality Metadata for Linked Data Source Selection
The traditional Web is evolving into the Web of Data which consists of huge collections
of structured data over poorly controlled distributed data sources. Live
queries are needed to get current information out of this global data space. In live
query processing, source selection deserves attention since it allows us to identify the
sources which might likely contain the relevant data. The thesis proposes a source
selection technique in the context of live query processing on Linked Open Data,
which takes into account the context of the request and the quality of data contained in
the sources to enhance the relevance (since the context enables a better interpretation
of the request) and the quality of the answers (which will be obtained by processing
the request on the selected sources). Specifically, the thesis proposes an extension of
the QTree indexing structure that had been proposed as a data summary to support
source selection based on source content, to take into account quality and contextual
information. With reference to a specific case study, the thesis also contributes an approach,
relying on the Luzzu framework, to assess the quality of a source with respect
to for a given context (according to different quality dimensions). An experimental
evaluation of the proposed techniques is also provide
Towards Querying in Decentralized Environments with Privacy-Preserving Aggregation
The Web is a ubiquitous economic, educational, and collaborative space.
However, it also serves as a haven for personal information harvesting.
Existing decentralised Web-based ecosystems, such as Solid, aim to combat
personal data exploitation on the Web by enabling individuals to manage their
data in the personal data store of their choice. Since personal data in these
decentralised ecosystems are distributed across many sources, there is a need
for techniques to support efficient privacy-preserving query execution over
personal data stores. Towards this end, in this position paper we present a
framework for efficient privacy preserving federated querying, and highlight
open research challenges and opportunities. The overarching goal being to
provide a means to position future research into privacy-preserving querying
within decentralised environments
Explora : interactive querying of multidimensional data in the context of smart cities
Citizen engagement is one of the key factors for smart city initiatives to remain sustainable over time. This in turn entails providing citizens and other relevant stakeholders with the latest data and tools that enable them to derive insights that add value to their day-to-day life. The massive volume of data being constantly produced in these smart city environments makes satisfying this requirement particularly challenging. This paper introduces Explora, a generic framework for serving interactive low-latency requests, typical of visual exploratory applications on spatiotemporal data, which leverages the stream processing for deriving-on ingestion time-synopsis data structures that concisely capture the spatial and temporal trends and dynamics of the sensed variables and serve as compacted data sets to provide fast (approximate) answers to visual queries on smart city data. The experimental evaluation conducted on proof-of-concept implementations of Explora, based on traditional database and distributed data processing setups, accounts for a decrease of up to 2 orders of magnitude in query latency compared to queries running on the base raw data at the expense of less than 10% query accuracy and 30% data footprint. The implementation of the framework on real smart city data along with the obtained experimental results prove the feasibility of the proposed approach
Interactive visual exploration of a large spatio-temporal dataset: Reflections on a geovisualization mashup
Exploratory visual analysis is useful for the preliminary investigation of large structured, multifaceted spatio-temporal datasets. This process requires the selection and aggregation of records by time, space and attribute, the ability to transform data and the flexibility to apply appropriate visual encodings and interactions. We propose an approach inspired by geographical 'mashups' in which freely-available functionality and data are loosely but flexibly combined using de facto exchange standards. Our case study combines MySQL, PHP and the LandSerf GIS to allow Google Earth to be used for visual synthesis and interaction with encodings described in KML. This approach is applied to the exploration of a log of 1.42 million requests made of a mobile directory service. Novel combinations of interaction and visual encoding are developed including spatial 'tag clouds', 'tag maps', 'data dials' and multi-scale density surfaces. Four aspects of the approach are informally evaluated: the visual encodings employed, their success in the visual exploration of the clataset, the specific tools used and the 'rnashup' approach. Preliminary findings will be beneficial to others considering using mashups for visualization. The specific techniques developed may be more widely applied to offer insights into the structure of multifarious spatio-temporal data of the type explored here
- …