3,425 research outputs found

    Impliance: A Next Generation Information Management Appliance

    Full text link
    ably successful in building a large market and adapting to the changes of the last three decades, its impact on the broader market of information management is surprisingly limited. If we were to design an information management system from scratch, based upon today's requirements and hardware capabilities, would it look anything like today's database systems?" In this paper, we introduce Impliance, a next-generation information management system consisting of hardware and software components integrated to form an easy-to-administer appliance that can store, retrieve, and analyze all types of structured, semi-structured, and unstructured information. We first summarize the trends that will shape information management for the foreseeable future. Those trends imply three major requirements for Impliance: (1) to be able to store, manage, and uniformly query all data, not just structured records; (2) to be able to scale out as the volume of this data grows; and (3) to be simple and robust in operation. We then describe four key ideas that are uniquely combined in Impliance to address these requirements, namely the ideas of: (a) integrating software and off-the-shelf hardware into a generic information appliance; (b) automatically discovering, organizing, and managing all data - unstructured as well as structured - in a uniform way; (c) achieving scale-out by exploiting simple, massive parallel processing, and (d) virtualizing compute and storage resources to unify, simplify, and streamline the management of Impliance. Impliance is an ambitious, long-term effort to define simpler, more robust, and more scalable information systems for tomorrow's enterprises.Comment: This article is published under a Creative Commons License Agreement (http://creativecommons.org/licenses/by/2.5/.) You may copy, distribute, display, and perform the work, make derivative works and make commercial use of the work, but, you must attribute the work to the author and CIDR 2007. 3rd Biennial Conference on Innovative Data Systems Research (CIDR) January 710, 2007, Asilomar, California, US

    Agent-Based Demand-Modeling Framework for Large-Scale Microsimulations

    Get PDF
    Microsimulation is becoming increasingly important in traffic demand modeling. The major advantage over traditional four-step models is the ability to simulate each traveler individually. Decision-making processes can be included for each individual. Traffic demand is the result of the different decisions made by individuals; these decisions lead to plans that the individuals then try to optimize. Therefore, such microsimulation models need appropriate initial demand patterns for all given individuals. The challenge is to create individual demand patterns out of general input data. In practice, there is a large variety of input data, which can differ in quality, spatial resolution, purpose, and other characteristics. The challenge for a flexible demand-modeling framework is to combine the various data types to produce individual demand patterns. In addition, the modeling framework has to define precise interfaces to provide portability to other models, programs, and frameworks, and it should be suitable for large-scale applications that use many millions of individuals. Because the model has to be adaptable to the given input data, the framework needs to be easily extensible with new algorithms and models. The presented demand-modeling framework for large-scale scenarios fulfils all these requirements. By modeling the demand for two different scenarios (Zurich, Switzerland, and the German states of Berlin and Brandenburg), the framework shows its flexibility in aspects of diverse input data, interfaces to third-party products, spatial resolution, and last but not least, the modeling process itself

    Information Integration - the process of integration, evolution and versioning

    Get PDF
    At present, many information sources are available wherever you are. Most of the time, the information needed is spread across several of those information sources. Gathering this information is a tedious and time consuming job. Automating this process would assist the user in its task. Integration of the information sources provides a global information source with all information needed present. All of these information sources also change over time. With each change of the information source, the schema of this source can be changed as well. The data contained in the information source, however, cannot be changed every time, due to the huge amount of data that would have to be converted in order to conform to the most recent schema.\ud In this report we describe the current methods to information integration, evolution and versioning. We distinguish between integration of schemas and integration of the actual data. We also show some key issues when integrating XML data sources

    The Relationship between Fuzzy Reasoning and Its Temporal Characteristics for Knowledge Management

    Get PDF
    The knowledge management systems based on artificial reasoning (KMAR) tries to provide computers the capabilities of performing various intelligent tasks for which their human users resort to their knowledge and collective intelligence. There is a need for incorporating aspects of time and imprecision into knowledge management systems, considering appropriate semantic foundations. The aim of this paper is to present the FRTES, a real-time fuzzy expert system, embedded in a knowledge management system. Our expert system is a special possibilistic expert system, developed in order to focus on fuzzy knowledge.Knowledge Management, Artificial Reasoning, predictability

    Informatics: the fuel for pharmacometric analysis

    Get PDF
    The current informal practice of pharmacometrics as a combination art and science makes it hard to appreciate the role that informatics can and should play in the future of the discipline and to comprehend the gaps that exist because of its absence. The development of pharmacometric informatics has important implications for expediting decision making and for improving the reliability of decisions made in model-based development. We argue that well-defined informatics for pharmacometrics can lead to much needed improvements in the efficiency, effectiveness, and reliability of the pharmacometrics process. The purpose of this paper is to provide a description of the pervasive yet often poorly appreciated role of informatics in improving the process of data assembly, a critical task in the delivery of pharmacometric analysis results. First, we provide a brief description of the pharmacometric analysis process. Second, we describe the business processes required to create analysis-ready data sets for the pharmacometrician. Third, we describe selected informatic elements required to support the pharmacometrics and data assembly processes. Finally, we offer specific suggestions for performing a systematic analysis of existing challenges as an approach to defi ning the next generation of pharmacometric informatics

    Design and implementation of an integrated surface texture information system for design, manufacture and measurement

    Get PDF
    The optimised design and reliable measurement of surface texture are essential to guarantee the functional performance of a geometric product. Current support tools are however often limited in functionality, integrity and efficiency. In this paper, an integrated surface texture information system for design, manufacture and measurement, called “CatSurf”, has been designed and developed, which aims to facilitate rapid and flexible manufacturing requirements. A category theory based knowledge acquisition and knowledge representation mechanism has been devised to retrieve and organize knowledge from various Geometrical Product Specifications (GPS) documents in surface texture. Two modules (for profile and areal surface texture) each with five components are developed in the CatSurf. It also focuses on integrating the surface texture information into a Computer-aided Technology (CAx) framework. Two test cases demonstrate design process of specifications for the profile and areal surface texture in AutoCAD and SolidWorks environments respectively

    UML-Based Approach to Developing Verified Embedded Software

    Get PDF

    Towards personalization in digital libraries through ontologies

    Get PDF
    In this paper we describe a browsing and searching personalization system for digital libraries based on the use of ontologies for describing the relationships between all the elements which take part in a digital library scenario of use. The main goal of this project is to help the users of a digital library to improve their experience of use by means of two complementary strategies: first, by maintaining a complete history record of his or her browsing and searching activities, which is part of a navigational user profile which includes preferences and all the aspects related to community involvement; and second, by reusing all the knowledge which has been extracted from previous usage from other users with similar profiles. This can be accomplished in terms of narrowing and focusing the search results and browsing options through the use of a recommendation system which organizes such results in the most appropriate manner, using ontologies and concepts drawn from the semantic web field. The complete integration of the experience of use of a digital library in the learning process is also pursued. Both the usage and information organization can be also exploited to extract useful knowledge from the way users interact with a digital library, knowledge that can be used to improve several design aspects of the library, ranging from internal organization aspects to human factors and user interfaces. Although this project is still on an early development stage, it is possible to identify all the desired functionalities and requirements that are necessary to fully integrate the use of a digital library in an e-learning environment
    • 

    corecore