45,390 research outputs found

    Knowledge management support for enterprise distributed systems

    No full text
    Explosion of information and increasing demands on semantic processing web applications have software systems to their limits. To address the problem we propose a semantic based formal framework (ADP) that makes use of promising technologies to enable knowledge generation and retrieval. We argue that this approach is cost effective, as it reuses and builds on existing knowledge and structure. It is also a good starting point for creating an organisational memory and providing knowledge management functions

    A quantitative analysis of the impact of a computerised information system on nurses' clinical practice using a realistic evaluation framework

    Get PDF
    Objective: To explore nurses' perceptions of the impact on clinical practice of the use of a computerised hospital information system. Design: A realistic evaluation design based on Pawson and Tilley's work has been used across all the phases of the study. This is a theory-driven approach and focuses evaluation on the study of what works, for whom and in what circumstances. These relationships are constructed as context-mechanisms-outcomes (CMO) configurations. Measurements: A questionnaire was distributed to all nurses working in in-patient units of a university hospital in Spain (n = 227). Quantitative data were analysed using SPSS 13.0. Descriptive statistics were used for an overall overview of nurses' perception. Inferential analysis, including both bivariate and multivariate methods (path analysis), was used for cross-tabulation of variables searching for CMO relationships. Results: Nurses (n = 179) participated in the study (78.8% response rate). Overall satisfaction with the IT system was positive. Comparisons with context variables show how nursing units' context had greater influence on perceptions than users' characteristics. Path analysis illustrated that the influence of unit context variables are on outcomes and not on mechanisms. Conclusion: Results from the study looking at subtle variations in users and units provide insight into how important professional culture and working practices could be in IT (information technology) implementation. The socio-technical approach on IT systems evaluation suggested in the recent literature appears to be an adequate theoretical underpinning for IT evaluation research. Realistic evaluation has proven to be an adequate method for IT evaluation. (C) 2009 Elsevier Ireland Ltd. All rights reserved

    Combined automotive safety and security pattern engineering approach

    Get PDF
    Automotive systems will exhibit increased levels of automation as well as ever tighter integration with other vehicles, traffic infrastructure, and cloud services. From safety perspective, this can be perceived as boon or bane - it greatly increases complexity and uncertainty, but at the same time opens up new opportunities for realizing innovative safety functions. Moreover, cybersecurity becomes important as additional concern because attacks are now much more likely and severe. However, there is a lack of experience with security concerns in context of safety engineering in general and in automotive safety departments in particular. To address this problem, we propose a systematic pattern-based approach that interlinks safety and security patterns and provides guidance with respect to selection and combination of both types of patterns in context of system engineering. A combined safety and security pattern engineering workflow is proposed to provide systematic guidance to support non-expert engineers based on best practices. The application of the approach is shown and demonstrated by an automotive case study and different use case scenarios.EC/H2020/692474/EU/Architecture-driven, Multi-concern and Seamless Assurance and Certification of Cyber-Physical Systems/AMASSEC/H2020/737422/EU/Secure COnnected Trustable Things/SCOTTEC/H2020/732242/EU/Dependability Engineering Innovation for CPS - DEIS/DEISBMBF, 01IS16043, Collaborative Embedded Systems (CrESt

    On systematic approaches for interpreted information transfer of inspection data from bridge models to structural analysis

    Get PDF
    In conjunction with the improved methods of monitoring damage and degradation processes, the interest in reliability assessment of reinforced concrete bridges is increasing in recent years. Automated imagebased inspections of the structural surface provide valuable data to extract quantitative information about deteriorations, such as crack patterns. However, the knowledge gain results from processing this information in a structural context, i.e. relating the damage artifacts to building components. This way, transformation to structural analysis is enabled. This approach sets two further requirements: availability of structural bridge information and a standardized storage for interoperability with subsequent analysis tools. Since the involved large datasets are only efficiently processed in an automated manner, the implementation of the complete workflow from damage and building data to structural analysis is targeted in this work. First, domain concepts are derived from the back-end tasks: structural analysis, damage modeling, and life-cycle assessment. The common interoperability format, the Industry Foundation Class (IFC), and processes in these domains are further assessed. The need for usercontrolled interpretation steps is identified and the developed prototype thus allows interaction at subsequent model stages. The latter has the advantage that interpretation steps can be individually separated into either a structural analysis or a damage information model or a combination of both. This approach to damage information processing from the perspective of structural analysis is then validated in different case studies

    Risk mitigation decisions for it security

    Get PDF
    Enterprises must manage their information risk as part of their larger operational risk management program. Managers must choose how to control for such information risk. This article defines the flow risk reduction problem and presents a formal model using a workflow framework. Three different control placement methods are introduced to solve the problem, and a comparative analysis is presented using a robust test set of 162 simulations. One year of simulated attacks is used to validate the quality of the solutions. We find that the math programming control placement method yields substantial improvements in terms of risk reduction and risk reduction on investment when compared to heuristics that would typically be used by managers to solve the problem. The contribution of this research is to provide managers with methods to substantially reduce information and security risks, while obtaining significantly better returns on their security investments. By using a workflow approach to control placement, which guides the manager to examine the entire infrastructure in a holistic manner, this research is unique in that it enables information risk to be examined strategically. © 2014 ACM

    Foggy clouds and cloudy fogs: a real need for coordinated management of fog-to-cloud computing systems

    Get PDF
    The recent advances in cloud services technology are fueling a plethora of information technology innovation, including networking, storage, and computing. Today, various flavors have evolved of IoT, cloud computing, and so-called fog computing, a concept referring to capabilities of edge devices and users' clients to compute, store, and exchange data among each other and with the cloud. Although the rapid pace of this evolution was not easily foreseeable, today each piece of it facilitates and enables the deployment of what we commonly refer to as a smart scenario, including smart cities, smart transportation, and smart homes. As most current cloud, fog, and network services run simultaneously in each scenario, we observe that we are at the dawn of what may be the next big step in the cloud computing and networking evolution, whereby services might be executed at the network edge, both in parallel and in a coordinated fashion, as well as supported by the unstoppable technology evolution. As edge devices become richer in functionality and smarter, embedding capacities such as storage or processing, as well as new functionalities, such as decision making, data collection, forwarding, and sharing, a real need is emerging for coordinated management of fog-to-cloud (F2C) computing systems. This article introduces a layered F2C architecture, its benefits and strengths, as well as the arising open and research challenges, making the case for the real need for their coordinated management. Our architecture, the illustrative use case presented, and a comparative performance analysis, albeit conceptual, all clearly show the way forward toward a new IoT scenario with a set of existing and unforeseen services provided on highly distributed and dynamic compute, storage, and networking resources, bringing together heterogeneous and commodity edge devices, emerging fogs, as well as conventional clouds.Peer ReviewedPostprint (author's final draft

    BIM adoption and implementation for architectural practices

    Get PDF
    Severe issues about data acquisition and management arise during the design creation and development due to complexity, uncertainty and ambiguity. BIM (Building Information Modelling) is a tool for a team based lean design approach towards improved architectural practice across the supply chain. However, moving from a CAD (Computer Aided Design) approach to BIM (Building Information Modelling) represents a fundamental change for individual disciplines and the construction industry as a whole. Although BIM has been implemented by large practices, it is not widely used by SMEs (Small and Medium Sized Enterprises). Purpose: This paper aims to present a systematic approach for BIM implementation for Architectural SMEs at the organizational level Design/Methodology/Approach: The research is undertaken through a KTP (Knowledge transfer Partnership) project between the University of Salford and John McCall Architects (JMA) a SME based in Liverpool. The overall aim of the KTP is to develop lean design practice through BIM adoption. The BIM implementation approach uses a socio-technical view which does not only consider the implementation of technology but also considers the socio-cultural environment that provides the context for its implementation. The action research oriented qualitative and quantitative research is used for discovery, comparison, and experimentation as it provides �learning by doing�. Findings: The strategic approach to BIM adoption incorporated people, process and technology equally and led to capacity building through the improvements in process, technological infrastructure and upskilling of JMA staff to attain efficiency gains and competitive advantages. Originality/Value: This paper introduces a systematic approach for BIM adoption based on the action research philosophy and demonstrates a roadmap for BIM adoption at the operational level for SME companie

    Sciunits: Reusable Research Objects

    Full text link
    Science is conducted collaboratively, often requiring knowledge sharing about computational experiments. When experiments include only datasets, they can be shared using Uniform Resource Identifiers (URIs) or Digital Object Identifiers (DOIs). An experiment, however, seldom includes only datasets, but more often includes software, its past execution, provenance, and associated documentation. The Research Object has recently emerged as a comprehensive and systematic method for aggregation and identification of diverse elements of computational experiments. While a necessary method, mere aggregation is not sufficient for the sharing of computational experiments. Other users must be able to easily recompute on these shared research objects. In this paper, we present the sciunit, a reusable research object in which aggregated content is recomputable. We describe a Git-like client that efficiently creates, stores, and repeats sciunits. We show through analysis that sciunits repeat computational experiments with minimal storage and processing overhead. Finally, we provide an overview of sharing and reproducible cyberinfrastructure based on sciunits gaining adoption in the domain of geosciences
    • …
    corecore