60,498 research outputs found

    Natural Language Query Processing for Model Management

    Get PDF
    The communication between an MIS and 1ts users would be greatly facilitated if the users could query and instruct the system in a sufficiently large subset of their natural language that the system appears to be conversing in the\u27 1 anguage. In large measure this has been accomplished for MISs that retrieve and display stored data and that perform simple calcul ations (summations, plots, regressions) with the data. A number of natural 1 anguage database query sys-· terns have been developed, a few of which are now commerci ally available. However, little attention has been , pald to the development of natural language interfaces for systems containing decision models. This paper examines the issues that may arise in the development of natural language , query processors for model management systems. In this paper we address four topics. The first ls the state-of-the-art in natural language database query processing. The principal issues here are the parsing of sentences and the resolution of ambiguities. The ambiguities may be those internal to a sentence (such as misspelllngs, ambiguities in the meanings of words, and ambiguities inherent in the syntax of the language in which the query is written), ambiguities resulting from explicit or implicit reference to previous queries (such as the use of pronouns whose referents must be identified), and ambiguities that arise when several flles must be combined to respond to a single query. These issues have been examined in detail and may provide a foundation for natural language model query processing. The second issue is the development of a high-level target language -- a well-structured, user-friendly, machineindependent language into which natural language queries will be translated prior to model execution. A target language for model management, called Mal (Model Query Language), has been designed, and its linguistic properties have been investigated. The language is described, and some exampl es are given. The thi rd issue is the structure of the model query transl ator. The transl ator w111 consist of five components. The Parsing Component receives the query from the user and analyzes it. It identifies the functions to be performed (e.g., optimization, sensitivity analysis), identifies the inputs and outputs of the models to be used, and attempts to resolve ambiguities. The Model Definition Component is used by the model buil der to define the inpu© and outputs of the model s l n the model bank. The Memory r_Aponent contai ns the model definitions, any previous queries (ln case reference, such as pronoun reference, is made to them), and information about possibl e spel 1 ing errors and synonyms. The Model Processing Component executes the model or model s needed to prepare a response, and the Report Writing Component formats the response. The final issue is the possible integration of data management and model management in a way that allows users to enter a natural 1 anguage query that requires access to both databases and model banks. Such an integration may eventual ly 1 ead to the devel opment of systems that provide a comprehensive range of decision support services.

    IMAGINE Final Report

    No full text

    Social media analytics: a survey of techniques, tools and platforms

    Get PDF
    This paper is written for (social science) researchers seeking to analyze the wealth of social media now available. It presents a comprehensive review of software tools for social networking media, wikis, really simple syndication feeds, blogs, newsgroups, chat and news feeds. For completeness, it also includes introductions to social media scraping, storage, data cleaning and sentiment analysis. Although principally a review, the paper also provides a methodology and a critique of social media tools. Analyzing social media, in particular Twitter feeds for sentiment analysis, has become a major research and business activity due to the availability of web-based application programming interfaces (APIs) provided by Twitter, Facebook and News services. This has led to an ‘explosion’ of data services, software tools for scraping and analysis and social media analytics platforms. It is also a research area undergoing rapid change and evolution due to commercial pressures and the potential for using social media data for computational (social science) research. Using a simple taxonomy, this paper provides a review of leading software tools and how to use them to scrape, cleanse and analyze the spectrum of social media. In addition, it discussed the requirement of an experimental computational environment for social media research and presents as an illustration the system architecture of a social media (analytics) platform built by University College London. The principal contribution of this paper is to provide an overview (including code fragments) for scientists seeking to utilize social media scraping and analytics either in their research or business. The data retrieval techniques that are presented in this paper are valid at the time of writing this paper (June 2014), but they are subject to change since social media data scraping APIs are rapidly changing

    My private cloud--granting federated access to cloud resources

    Get PDF
    We describe the research undertaken in the six month JISC/EPSRC funded My Private Cloud project, in which we built a demonstration cloud file storage service that allows users to login to it, by using their existing credentials from a configured trusted identity provider. Once authenticated, users are shown a set of accounts that they are the owners of, based on their identity attributes. Once users open one of their accounts, they can upload and download files to it. Not only that, but they can then grant access to their file resources to anyone else in the federated system, regardless of whether their chosen delegate has used the cloud service before or not. The system uses standard identity management protocols, attribute based access controls, and a delegation service. A set of APIs have been defined for the authentication, authorisation and delegation processes, and the software has been released as open source to the community. A public demonstration of the system is available online

    The Semantic Automated Discovery and Integration (SADI) Web service Design-Pattern, API and Reference Implementation

    Get PDF
    Background. 
The complexity and inter-related nature of biological data poses a difficult challenge for data and tool integration. There has been a proliferation of interoperability standards and projects over the past decade, none of which has been widely adopted by the bioinformatics community. Recent attempts have focused on the use of semantics to assist integration, and Semantic Web technologies are being welcomed by this community.

Description. 
SADI – Semantic Automated Discovery and Integration – is a lightweight set of fully standards-compliant Semantic Web service design patterns that simplify the publication of services of the type commonly found in bioinformatics and other scientific domains. Using Semantic Web technologies at every level of the Web services “stack”, SADI services consume and produce instances of OWL Classes following a small number of very straightforward best-practices. In addition, we provide codebases that support these best-practices, and plug-in tools to popular developer and client software that dramatically simplify deployment of services by providers, and the discovery and utilization of those services by their consumers.

Conclusions.
SADI Services are fully compliant with, and utilize only foundational Web standards; are simple to create and maintain for service providers; and can be discovered and utilized in a very intuitive way by biologist end-users. In addition, the SADI design patterns significantly improve the ability of software to automatically discover appropriate services based on user-needs, and automatically chain these into complex analytical workflows. We show that, when resources are exposed through SADI, data compliant with a given ontological model can be automatically gathered, or generated, from these distributed, non-coordinating resources - a behavior we have not observed in any other Semantic system. Finally, we show that, using SADI, data dynamically generated from Web services can be explored in a manner very similar to data housed in static triple-stores, thus facilitating the intersection of Web services and Semantic Web technologies

    From Questions to Effective Answers: On the Utility of Knowledge-Driven Querying Systems for Life Sciences Data

    Get PDF
    We compare two distinct approaches for querying data in the context of the life sciences. The first approach utilizes conventional databases to store the data and intuitive form-based interfaces to facilitate easy querying of the data. These interfaces could be seen as implementing a set of "pre-canned" queries commonly used by the life science researchers that we study. The second approach is based on semantic Web technologies and is knowledge (model) driven. It utilizes a large OWL ontology and same datasets as before but associated as RDF instances of the ontology concepts. An intuitive interface is provided that allows the formulation of RDF triples-based queries. Both these approaches are being used in parallel by a team of cell biologists in their daily research activities, with the objective of gradually replacing the conventional approach with the knowledge-driven one. This provides us with a valuable opportunity to compare and qualitatively evaluate the two approaches. We describe several benefits of the knowledge-driven approach in comparison to the traditional way of accessing data, and highlight a few limitations as well. We believe that our analysis not only explicitly highlights the specific benefits and limitations of semantic Web technologies in our context but also contributes toward effective ways of translating a question in a researcher's mind into precise computational queries with the intent of obtaining effective answers from the data. While researchers often assume the benefits of semantic Web technologies, we explicitly illustrate these in practice
    • 

    corecore