5,251 research outputs found

    Exact two-terminal reliability of some directed networks

    Full text link
    The calculation of network reliability in a probabilistic context has long been an issue of practical and academic importance. Conventional approaches (determination of bounds, sums of disjoint products algorithms, Monte Carlo evaluations, studies of the reliability polynomials, etc.) only provide approximations when the network's size increases, even when nodes do not fail and all edges have the same reliability p. We consider here a directed, generic graph of arbitrary size mimicking real-life long-haul communication networks, and give the exact, analytical solution for the two-terminal reliability. This solution involves a product of transfer matrices, in which individual reliabilities of edges and nodes are taken into account. The special case of identical edge and node reliabilities (p and rho, respectively) is addressed. We consider a case study based on a commonly-used configuration, and assess the influence of the edges being directed (or not) on various measures of network performance. While the two-terminal reliability, the failure frequency and the failure rate of the connection are quite similar, the locations of complex zeros of the two-terminal reliability polynomials exhibit strong differences, and various structure transitions at specific values of rho. The present work could be extended to provide a catalog of exactly solvable networks in terms of reliability, which could be useful as building blocks for new and improved bounds, as well as benchmarks, in the general case

    Experiments in fault tolerant software reliability

    Get PDF
    The reliability of voting was evaluated in a fault-tolerant software system for small output spaces. The effectiveness of the back-to-back testing process was investigated. Version 3.0 of the RSDIMU-ATS, a semi-automated test bed for certification testing of RSDIMU software, was prepared and distributed. Software reliability estimation methods based on non-random sampling are being studied. The investigation of existing fault-tolerance models was continued and formulation of new models was initiated

    Reliability models for dataflow computer systems

    Get PDF
    The demands for concurrent operation within a computer system and the representation of parallelism in programming languages have yielded a new form of program representation known as data flow (DENN 74, DENN 75, TREL 82a). A new model based on data flow principles for parallel computations and parallel computer systems is presented. Necessary conditions for liveness and deadlock freeness in data flow graphs are derived. The data flow graph is used as a model to represent asynchronous concurrent computer architectures including data flow computers

    An experiment in software reliability

    Get PDF
    The results of a software reliability experiment conducted in a controlled laboratory setting are reported. The experiment was undertaken to gather data on software failures and is one in a series of experiments being pursued by the Fault Tolerant Systems Branch of NASA Langley Research Center to find a means of credibly performing reliability evaluations of flight control software. The experiment tests a small sample of implementations of radar tracking software having ultra-reliability requirements and uses n-version programming for error detection, and repetitive run modeling for failure and fault rate estimation. The experiment results agree with those of Nagel and Skrivan in that the program error rates suggest an approximate log-linear pattern and the individual faults occurred with significantly different error rates. Additional analysis of the experimental data raises new questions concerning the phenomenon of interacting faults. This phenomenon may provide one explanation for software reliability decay

    Optimal Allocation of Resources in Reliability Growth

    Get PDF
    Reliability growth testing seeks to identify and remove failure modes in order to improve system reliability. This dissertation centers around the resource allocation across the components of a multi-component system to maximize system reliability. We summarize this dissertation’s contributions to optimal resource allocation in reliability growth. Chapter 2 seeks to deploy limited testing resources across the components of a series-parallel system in effort to maximize system reliability under the assumption that each component’s reliability exhibits growth according to an AMSAA model with known parameters. An optimization model for this problem is developed and then extended to consider the allocation of testing resources in a series-parallel system with consideration for the possibility of testing at different levels (system, subsystem, and component). We contribute a class of exact algorithms that decomposes the problem based upon the series-parallel structure. We prove the algorithm is finite, compare it with heuristic approaches on a set of test instances, and provide detailed analyses of numerical examples. In Chapter 3, we extend model in Chapter 2 to solve a robust optimization version of this problem in which AMSAA parameters are uncertain but assumed to lie within a budget-restricted uncertainty set. We model the problem of robust allocation of testing resources to maximize system reliability for both series and series-parallel systems, and we develop and analyze exact solution approaches for this problem based on a cutting plane algorithm. Computational results demonstrate the value of the robust optimization approach as compared to deterministic alternatives. In the last chapter, we develop a new model that merges testing components and installing redundancies within an integrated optimization model that maximizes system reliability. Specifically, our model considers a series-parallel system in which the system reliability can be improved by both testing components and installing redundant components. We contribute an exact algorithm that decomposes the problem into smaller integer linear programs. We prove that this algorithm is finite and apply it to a set of instances. Experiments demonstrate that the integrated approach generates greater reliabilities than applying test planning and redundancy allocation models iteratively, and moreover, it yields significant savings in computational time

    Conceptual design and feasibility evaluation model of a 10 to the 8th power bit oligatomic mass memory. Volume 1: Conceptual design

    Get PDF
    The oligatomic (mirror) thin film memory technology is a suitable candidate for general purpose spaceborne applications in the post-1975 time frame. Capacities of around 10 to the 8th power bits can be reliably implemented with systems designed around a 335 million bit module. The recommended mode was determined following an investigation of implementation sizes ranging from an 8,000,000 to 100,000,000 bits per module. Cost, power, weight, volume, reliability, maintainability and speed were investigated. The memory includes random access, NDRO, SEC-DED, nonvolatility, and dual interface characteristics. The applications most suitable for the technology are those involving a large capacity with high speed (no latency), nonvolatility, and random accessing

    An approach to software reliability prediction and quality control

    Get PDF
    December 5-7, 1972, Fall Joint Computer ConferenceThe increase in importance of software in command and control and other complex systems has not been accompanied by commensurate progress in the develop- ment of analytical techniques for the measurement of software quality and the prediction of software reliability. This paper presents a rationale for imple- menting software reliability programs; defines software reliability; and describes some of the problems of performing software reliability analysis. A software reliability program is outlined and a methodology for reliability prediction and quality control is presented. The results of initial efforts to develop a software reliability methodology at the Naval Electronics Laboratory Center are reported

    Users manual for the Automated Performance Test System (APTS)

    Get PDF
    The characteristics of and the user information for the Essex Automated Performance Test System (APTS) computer-based portable performance assessment battery are given. The battery was developed to provide a menu of performance test tapping the widest possible variety of human cognitive and motor functions, implemented on a portable computer system suitable for use in both laboratory and field settings for studying the effects of toxic agents and other stressors. The manual gives guidance in selecting, administering and scoring tests from the battery, and reviews the data and studies underlying the development of the battery. Its main emphasis is on the users of the battery - the scientists, researchers and technicians who wish to examine changes in human performance across time or as a function of changes in the conditions under which test data are obtained. First the how to information needed to make decisions about where and how to use the battery is given, followed by the research background supporting the battery development. Further, the development history of the battery focuses largely on the logical framework within which tests were evaluated

    A contingency model of perceived effectiveness in accounting information systems: Organizational coordination and control effects

    Get PDF
    A contingency model is advanced that examines sources of requirements for organizational coordination and control as they affect the extent of integration in an accounting information system. Requirements that are contingent on the degree of organizational formalization, information interdependence among functional areas, and dependence in interorganizational information sharing and electronic data interchange links, are examined. The congruence or fit of system integration with those requirements is a key concept that influences beliefs about system effectiveness. Results of the empirical study indicated that, as hypothesized, the fit between the accounting system design and the contingency factors resulted in a more successful system. Specifically, system fit was a significant factor that explained variations in perceived AIS effectiveness, as measured by decision makers’ perceived satisfaction with the accuracy and monitoring effectiveness of output information. The effect of system fit on a second factor of perceived AIS effectiveness, as measured by decision-makers’ satisfaction with the perceived quality of information content in system outputs, was only marginally significant. The study addresses an important area in accounting systems research that directly relates to the decision facilitation and control objectives of accounting information. © 2000 Elsevier Science Inc. All rights reserved
    • …
    corecore