3,424 research outputs found
Realization of Semantic Atom Blog
Web blog is used as a collaborative platform to publish and share
information. The information accumulated in the blog intrinsically contains the
knowledge. The knowledge shared by the community of people has intangible value
proposition. The blog is viewed as a multimedia information resource available
on the Internet. In a blog, information in the form of text, image, audio and
video builds up exponentially. The multimedia information contained in an Atom
blog does not have the capability, which is required by the software processes
so that Atom blog content can be accessed, processed and reused over the
Internet. This shortcoming is addressed by exploring OWL knowledge modeling,
semantic annotation and semantic categorization techniques in an Atom blog
sphere. By adopting these techniques, futuristic Atom blogs can be created and
deployed over the Internet
BlogForever D2.4: Weblog spider prototype and associated methodology
The purpose of this document is to present the evaluation of different solutions for capturing blogs, established methodology and to describe the developed blog spider prototype
BlogForever D2.6: Data Extraction Methodology
This report outlines an inquiry into the area of web data extraction, conducted within the context of blog preservation. The report reviews theoretical advances and practical developments for implementing data extraction. The inquiry is extended through an experiment that demonstrates the effectiveness and feasibility of implementing some of the suggested approaches. More specifically, the report discusses an approach based on unsupervised machine learning that employs the RSS feeds and HTML representations of blogs. It outlines the possibilities of extracting semantics available in blogs and demonstrates the benefits of exploiting available standards such as microformats and microdata. The report proceeds to propose a methodology for extracting and processing blog data to further inform the design and development of the BlogForever platform
Recommended from our members
Consuming Linked Data within a Large Educational Organization
Large universities tend to spread their services across several departments to serve their substantial student base. It is very common for this to result in developing different systems, which end up in creating many disconnected data silos within the organization. Data isolation is one of the main bottlenecks that prevent unlocking the full potential behind exploiting such data, to provide a better experience at the level of application deployment and data analysis. The Open University is in the process of connecting their data silos by relying on the Linked Data principles within the LUCERO project. We discuss in this paper three use-cases through which we consume Linked Data produced at the Open University: (1) a student services use-case showing how we exploit data connections to deliver learning material related to courses through the university's main course information website; (2) a mobile course application that enables students to easily explore courses by subject, qualification or research topic; and (3) a Leanback TV service that provides students the ability to watch, with a degree of control, a set of podcasts grouped in different channels. Through these use cases, we highlight in this paper the advantages and effects of consuming Linked Data within an organization
Social media analytics: a survey of techniques, tools and platforms
This paper is written for (social science) researchers seeking to analyze the wealth of social media now available. It presents a comprehensive review of software tools for social networking media, wikis, really simple syndication feeds, blogs, newsgroups, chat and news feeds. For completeness, it also includes introductions to social media scraping, storage, data cleaning and sentiment analysis. Although principally a review, the paper also provides a methodology and a critique of social media tools. Analyzing social media, in particular Twitter feeds for sentiment analysis, has become a major research and business activity due to the availability of web-based application programming interfaces (APIs) provided by Twitter, Facebook and News services. This has led to an ‘explosion’ of data services, software tools for scraping and analysis and social media analytics platforms. It is also a research area undergoing rapid change and evolution due to commercial pressures and the potential for using social media data for computational (social science) research. Using a simple taxonomy, this paper provides a review of leading software tools and how to use them to scrape, cleanse and analyze the spectrum of social media. In addition, it discussed the requirement of an experimental computational environment for social media research and presents as an illustration the system architecture of a social media (analytics) platform built by University College London. The principal contribution of this paper is to provide an overview (including code fragments) for scientists seeking to utilize social media scraping and analytics either in their research or business. The data retrieval techniques that are presented in this paper are valid at the time of writing this paper (June 2014), but they are subject to change since social media data scraping APIs are rapidly changing
Content Accessibility of Web documents. Principles and Recommendations
The paper is an overview of issues related to the accessibility of Web sites, of European initiatives and recommendations in the field, of future solutions, such as developing Web 2.0 accessible applications with WAI-ARIA. As part of the Pro-Inclusiv project was included a set of accessibility recommendations to design web sites, recommendations presented in the paper.Web accessibility, users with disabilities, standards and recommendations
Representation and use of chemistry in the global electronic age.
We present an overview of the current state of public semantic chemistry and propose new approaches at a strategic and a detailed level. We show by example how a model for a Chemical Semantic Web can be constructed using machine-processed data and information from journal articles.This manuscript addresses questions of robotic access to data and its automatic re-use, including the role of Open Access archival of data. This is a pre-refereed preprint allowed by the publisher's (Royal Soc. Chemistry) Green policy. The author's preferred manuscript is an HTML hyperdocument with ca. 20 links to images, some of which are JPEgs and some of which are SVG (scalable vector graphics) including animations. There are also links to molecules in CML, for which the Jmol viewer is recommended. We susgeest that readers who wish to see the full glory of the manuscript, download the Zipped version and unpack on their machine. We also supply a PDF and DOC (Word) version which obviously cannot show the animations, but which may be the best palce to start, particularly for those more interested in the text
Defining interoperability standards: A case study of public health observatory websites
The Association of Public Health Observatories (APHO) is a group of region-based health-information providers. Each PHO publishes health-related data for their specific region. Each observatory has taken a national lead in one or more key health area - such as 'cancer' or Obesity'. In 2003, a project was initiated to develop 'interoperability' between public health observatory websites, so the national resources published by one lead observatory could be found on the websites for each other PHO. The APHO interoperability project defined a set of requirements for each PHO - websites should comply with the current government data standards and provide webservices to allow data to be searched in real-time between different PHOs. This thesis describes the production of an interoperable website for the North East Public Health Observatory (NEPHO) and the problems faced during implementation to comply with the APHO interoperability requirements. The areas of interoperability, e-Government and metadata were investigated specifically in suitability for NEPHO and an action list of tasks necessary to achieve the project aims was drawn up. This project has resulted in the successful introduction of a new NEPHO website that complies with the APHO and e-Govemment requirements, however interoperability with other organisations has been difficult to achieve. This thesis describes how other organisations approached the same APHO interoperability criteria and questions whether the national project governance could be improved
- …