4,716 research outputs found

    Instant restore after a media failure

    Full text link
    Media failures usually leave database systems unavailable for several hours until recovery is complete, especially in applications with large devices and high transaction volume. Previous work introduced a technique called single-pass restore, which increases restore bandwidth and thus substantially decreases time to repair. Instant restore goes further as it permits read/write access to any data on a device undergoing restore--even data not yet restored--by restoring individual data segments on demand. Thus, the restore process is guided primarily by the needs of applications, and the observed mean time to repair is effectively reduced from several hours to a few seconds. This paper presents an implementation and evaluation of instant restore. The technique is incrementally implemented on a system starting with the traditional ARIES design for logging and recovery. Experiments show that the transaction latency perceived after a media failure can be cut down to less than a second and that the overhead imposed by the technique on normal processing is minimal. The net effect is that a few "nines" of availability are added to the system using simple and low-overhead software techniques

    Data center resilience assessment : storage, networking and security.

    Get PDF
    Data centers (DC) are the core of the national cyber infrastructure. With the incredible growth of critical data volumes in financial institutions, government organizations, and global companies, data centers are becoming larger and more distributed posing more challenges for operational continuity in the presence of experienced cyber attackers and occasional natural disasters. The main objective of this research work is to present a new methodology for data center resilience assessment, this methodology consists of: • Define Data center resilience requirements. • Devise a high level metric for data center resilience. • Design and develop a tool to validate and the metric. Since computer networks are an important component in the data center architecture, this research work was extended to investigate computer network resilience enhancement opportunities within the area of routing protocols, redundancy, and server load to minimize the network down time and increase the time period of resisting attacks. Data center resilience assessment is a complex process as it involves several aspects such as: policies for emergencies, recovery plans, variation in data center operational roles, hosted/processed data types and data center architectures. However, in this dissertation, storage, networking and security are emphasized. The need for resilience assessment emerged due to the gap in existing reliability, availability, and serviceability (RAS) measures. Resilience as an evaluation metric leads to better proactive perspective in system design and management. The proposed Data center resilience assessment portal (DC-RAP) is designed to easily integrate various operational scenarios. DC-RAP features a user friendly interface to assess the resilience in terms of performance analysis and speed recovery by collecting the following information: time to detect attacks, time to resist, time to fail and recovery time. Several set of experiments were performed, results obtained from investigating the impact of routing protocols, server load balancing algorithms on network resilience, showed that using particular routing protocol or server load balancing algorithm can enhance network resilience level in terms of minimizing the downtime and ensure speed recovery. Also experimental results for investigating the use social network analysis (SNA) for identifying important router in computer network showed that the SNA was successful in identifying important routers. This important router list can be used to redundant those routers to ensure high level of resilience. Finally, experimental results for testing and validating the data center resilience assessment methodology using the DC-RAP showed the ability of the methodology quantify data center resilience in terms of providing steady performance, minimal recovery time and maximum resistance-attacks time. The main contributions of this work can be summarized as follows: • A methodology for evaluation data center resilience has been developed. • Implemented a Data Center Resilience Assessment Portal (D$-RAP) for resilience evaluations. • Investigated the usage of Social Network Analysis to Improve the computer network resilience

    A Survey of Fault-Tolerance and Fault-Recovery Techniques in Parallel Systems

    Full text link
    Supercomputing systems today often come in the form of large numbers of commodity systems linked together into a computing cluster. These systems, like any distributed system, can have large numbers of independent hardware components cooperating or collaborating on a computation. Unfortunately, any of this vast number of components can fail at any time, resulting in potentially erroneous output. In order to improve the robustness of supercomputing applications in the presence of failures, many techniques have been developed to provide resilience to these kinds of system faults. This survey provides an overview of these various fault-tolerance techniques.Comment: 11 page

    A Taxonomy of Data Grids for Distributed Data Sharing, Management and Processing

    Full text link
    Data Grids have been adopted as the platform for scientific communities that need to share, access, transport, process and manage large data collections distributed worldwide. They combine high-end computing technologies with high-performance networking and wide-area storage management techniques. In this paper, we discuss the key concepts behind Data Grids and compare them with other data sharing and distribution paradigms such as content delivery networks, peer-to-peer networks and distributed databases. We then provide comprehensive taxonomies that cover various aspects of architecture, data transportation, data replication and resource allocation and scheduling. Finally, we map the proposed taxonomy to various Data Grid systems not only to validate the taxonomy but also to identify areas for future exploration. Through this taxonomy, we aim to categorise existing systems to better understand their goals and their methodology. This would help evaluate their applicability for solving similar problems. This taxonomy also provides a "gap analysis" of this area through which researchers can potentially identify new issues for investigation. Finally, we hope that the proposed taxonomy and mapping also helps to provide an easy way for new practitioners to understand this complex area of research.Comment: 46 pages, 16 figures, Technical Repor

    STRUCTURAL MODEL OF THE SYSTEM OF DEVELOPMENT OF METHODOLOGICAL COMPETENCE OF IT-DISCIPLINE TEACHERS ON THE BASIS OF CONTINUING EDUCATION

    Get PDF
    The article considers the creation of structural model for the system of development of methodological competence of IT-discipline teachers based on continuing education with the description of the structure of microservices subsystems. To implement the proposed conceptual model the paper proposes a software architecture based on the use of classical service-oriented design pattern "Model-View Separation" (separation of model objects from the graphical interface objects with the user) and provides for interaction with the database, the composition of its main packages is detailed. Logical and physical service-oriented models of basic classes of objects of subject area and auxiliary objects (services) have been developed. The article also proposes a relational database structure based on the use of metadata tables, directories, and operational data in a relational database management system environment. Methods and mechanisms for data protection and replication are proposed and justified, considering the specifics of system operation in higher education institutions. The article aims to develop insurgents and mechanisms of creation of information technology improvement of the system of higher education and creation and operation of integrated flexible software tools to support the mixed system of organization of educational process, as well as increasing the effectiveness of the introduction of a mixed system for the organization of the educational process on the basis of the use of the latest technical means of training and documentation, ensuring free access of students to teaching and methodological resources, сontent and criteria for the assessment of knowledge in disciplines and modules, data recording, and partial compensation for the increased workload of teachers and staff through the automation of assessment, registration and assessment processes. A task-oriented model for managing technical problems in the implementation of the information system for the development of teachers' methodological competence is also proposed, and a conceptual framework for managing these problems has been developed from the perspective of considering projects as a set of objects and subjects. Technical problems in projects are shown to be obstacles to moving in the right direction, created by random events, uncertainties, and changes. A mathematical model of the change in resistance value caused by technical problems has been developed and methods to counteract these problems have been proposed

    A Technology Proposal for a Management Information System for the Director’s Office, NAL.

    Get PDF
    This technology proposal attempts in giving a viable solution for a Management Information System (MIS) for the Director's Office. In today's IT scenario, an Organization's success greatly depends on its ability to get accurate and timely data on its operations of varied nature and to manage this data effectively to guide its activities and meet its goals. To cater to the information needs of an Organization or an Office like the Director's Office, information systems are developed and deployed to gather and process data in ways that produce a variety of information to the end-user. MIS can therefore can be defined as an integrated user-machine system for providing information to support operations, management and decision-making functions in an Organization. The system in a nutshell, utilizes computer hardware and software, manual procedures, models for analysis planning, control and decision-making and a database. Using state-of-the-art front-end and back-end web based tools, this technology proposal attempts to provide a single-point Information Management, Information Storage, Information Querying and Information Retrieval interface to the Director and his office for handling all information traffic flow in and out of the Director's Office

    A Study of Mongodb and Oracle in an E-Commerce Environment

    Get PDF
    As worldwide e-commerce expands, businesses continue to look for better ways to meet their evolving needs with web solutions that scale and perform adequately. Several online retailers have been able to address scaling challenges through the implementation of NoSQL databases. While architecturally different from their relational database counterparts, NoSQL databases typically achieve performance gains by relaxing one or more of the essential transaction processing attributes of atomicity, consistency, isolation, and durability (ACID). As with any emerging technology, there are both critics and supporters of NoSQL databases. The detractors claim that NoSQL is not safe and is at a greater risk for data loss. On the other hand, its ardent defenders boast the performance gains achieved over their relational counterparts. This thesis studies the NoSQL database known as MongoDB, and discusses its ability to support the growing needs of e-commerce data processing. It then examines MongoDB\u27s raw performance (compared to Oracle 11g R2, a relational database) and discusses its adherence to ACID
    corecore