1,465,135 research outputs found
An Infrastructure for acquiring high quality semantic metadata
Because metadata that underlies semantic web applications is gathered from distributed and heterogeneous data sources, it is important to ensure its quality (i.e., reduce duplicates, spelling errors, ambiguities). However, current infrastructures that acquire and integrate semantic data have only marginally addressed the issue of metadata quality. In this paper we present our metadata acquisition infrastructure, ASDI, which pays special attention to ensuring that high quality metadata is derived. Central to the architecture of ASDI is a erification engine that relies on several semantic web tools to check the quality of the derived data. We tested our prototype in the context of building a semantic web portal for our lab, KMi. An experimental evaluation omparing the automatically extracted data against manual annotations indicates that the verification engine enhances the quality of the extracted semantic metadata
Archiving the Relaxed Consistency Web
The historical, cultural, and intellectual importance of archiving the web
has been widely recognized. Today, all countries with high Internet penetration
rate have established high-profile archiving initiatives to crawl and archive
the fast-disappearing web content for long-term use. As web technologies
evolve, established web archiving techniques face challenges. This paper
focuses on the potential impact of the relaxed consistency web design on
crawler driven web archiving. Relaxed consistent websites may disseminate,
albeit ephemerally, inaccurate and even contradictory information. If captured
and preserved in the web archives as historical records, such information will
degrade the overall archival quality. To assess the extent of such quality
degradation, we build a simplified feed-following application and simulate its
operation with synthetic workloads. The results indicate that a non-trivial
portion of a relaxed consistency web archive may contain observable
inconsistency, and the inconsistency window may extend significantly longer
than that observed at the data store. We discuss the nature of such quality
degradation and propose a few possible remedies.Comment: 10 pages, 6 figures, CIKM 201
A walk through the webās video clips
Approximately 10^5 video clips are posted every day on the Web. The popularity of Web-based video databases poses a number of challenges to machine vision scientists: how do we organize, index and search such large wealth of data? Content-based video search and classification have been proposed in the literature and applied successfully to analyzing movies, TV broadcasts and lab-made videos. We explore the performance of some of these algorithms on a large data-set of approximately 3000 videos. We collected our data-set directly from the Web minimizing bias for content or quality, way so as to have a faithful representation of the statistics of this medium. We find that the algorithms that we have come to trust do not work well on video clips, because their quality is lower and their subject is more varied. We will make the data publicly available to encourage further research
Composing Distributed Data-intensive Web Services Using a Flexible Memetic Algorithm
Web Service Composition (WSC) is a particularly promising application of Web
services, where multiple individual services with specific functionalities are
composed to accomplish a more complex task, which must fulfil functional
requirements and optimise Quality of Service (QoS) attributes, simultaneously.
Additionally, large quantities of data, produced by technological advances,
need to be exchanged between services. Data-intensive Web services, which
manipulate and deal with those data, are of great interest to implement
data-intensive processes, such as distributed Data-intensive Web Service
Composition (DWSC). Researchers have proposed Evolutionary Computing (EC)
fully-automated WSC techniques that meet all the above factors. Some of these
works employed Memetic Algorithms (MAs) to enhance the performance of EC
through increasing its exploitation ability of in searching neighbourhood area
of a solution. However, those works are not efficient or effective. This paper
proposes an MA-based approach to solving the problem of distributed DWSC in an
effective and efficient manner. In particular, we develop an MA that hybridises
EC with a flexible local search technique incorporating distance of services.
An evaluation using benchmark datasets is carried out, comparing existing
state-of-the-art methods. Results show that our proposed method has the highest
quality and an acceptable execution time overall.Comment: arXiv admin note: text overlap with arXiv:1901.0556
Social Ranking Techniques for the Web
The proliferation of social media has the potential for changing the
structure and organization of the web. In the past, scientists have looked at
the web as a large connected component to understand how the topology of
hyperlinks correlates with the quality of information contained in the page and
they proposed techniques to rank information contained in web pages. We argue
that information from web pages and network data on social relationships can be
combined to create a personalized and socially connected web. In this paper, we
look at the web as a composition of two networks, one consisting of information
in web pages and the other of personal data shared on social media web sites.
Together, they allow us to analyze how social media tunnels the flow of
information from person to person and how to use the structure of the social
network to rank, deliver, and organize information specifically for each
individual user. We validate our social ranking concepts through a ranking
experiment conducted on web pages that users shared on Google Buzz and Twitter.Comment: 7 pages, ASONAM 201
Recommended from our members
Interactive Water Quality Data Visualization: Case Study Longhorn Stream Team
Water quality monitoring can provide a real-time indicator as to the health and quality of a precious natural resource yet it is difficult to effectively communicate data and inspire action beyond the scientific arena. An interactive web application was developed using R Shiny to bridge this knowledge gap, visualizing existing data points collected by citizen scientists of the Longhorn Stream Team.Geography and the Environmen
- ā¦