11,593 research outputs found
Recommended from our members
Evaluation of software dependability
It has been said that the term software engineering is an aspiration not a description. We would like to be able to claim that we engineer software, in the same sense that we engineer an aero-engine, but most of us would agree that this is not currently an accurate description of our activities. My suspicion is that it never will be.
From the point of view of this essay – i.e. dependability evaluation – a major difference between software and other engineering artefacts is that the former is pure design. Its unreliability is always the result of design faults, which in turn arise as a result of human intellectual failures. The unreliability of hardware systems, on the other hand, has tended until recently to be dominated by random physical failures of components – the consequences of the ‘perversity of nature’. Reliability theories have been developed over the years which have successfully allowed systems to be built to high reliability requirements, and the final system reliability to be evaluated accurately. Even for pure hardware systems, without software, however, the very success of these theories has more recently highlighted the importance of design faults in determining the overall reliability of the final product. The conventional hardware reliability theory does not address this problem at all.
In the case of software, there is no physical source of failures, and so none of the reliability theory developed for hardware is relevant. We need new theories that will allow us to achieve required dependability levels, and to evaluate the actual dependability that has been achieved, when the sources of the faults that ultimately result in failure are human intellectual failures
Reliability Assessment of a Packaging Automatic Machine by Accelerated Life Testing Approach
Industrial competitiveness in innovation, the time of the market introduction of new machines and the level of reliability requested implies that the strategies for the development of products must be more and more efficient. In particular, researchers and practitioners are looking for methods to evaluate the reliability, as cheap as possible, knowing that systems are more and more reliable. This paper presents a reliability assessment procedure applied to a mechanical component of an automatic machine for packaging using the accelerated test approach. The general log-linear (GLL) model is combined based on a relationship between a number strains, in particular mechanical and time based. The complete Accelerated Life Testing - ALT approach is presented by using Weibull distribution and Maximum Likelihood verifying method. A test plan is proposed to estimate the unknown parameters of accelerated life models. Using the proposed ALT model, the reliability function of the component is evaluated and then compared with data from the field collected by customers referring to 8 years of real work on a fleet of automatic packaging machines.
The results confirm that the assessment method through ALT is effective for lifetime prediction with shorter test times, and for the same reason it can improve the design process of automatic packaging machines
Validation of Ultrahigh Dependability for Software-Based Systems
Modern society depends on computers for a number of critical tasks in which failure can have very high costs. As a consequence, high levels of dependability (reliability, safety, etc.) are required from such computers, including their software. Whenever a quantitative approach to risk is adopted, these requirements must be stated in quantitative terms, and a rigorous demonstration of their being attained is necessary. For software used in the most critical roles, such demonstrations are not usually supplied. The fact is that the dependability requirements often lie near the limit of the current state of the art, or beyond, in terms not only of the ability to satisfy them, but also, and more often, of the ability to demonstrate that they are satisfied in the individual operational products (validation). We discuss reasons why such demonstrations cannot usually be provided with the means available: reliability growth models, testing with stable reliability, structural dependability modelling, as well as more informal arguments based on good engineering practice. We state some rigorous arguments about the limits of what can be validated with each of such means. Combining evidence from these different sources would seem to raise the levels that can be validated; yet this improvement is not such as to solve the problem. It appears that engineering practice must take into account the fact that no solution exists, at present, for the validation of ultra-high dependability in systems relying on complex software
Expert Elicitation for Reliable System Design
This paper reviews the role of expert judgement to support reliability
assessments within the systems engineering design process. Generic design
processes are described to give the context and a discussion is given about the
nature of the reliability assessments required in the different systems
engineering phases. It is argued that, as far as meeting reliability
requirements is concerned, the whole design process is more akin to a
statistical control process than to a straightforward statistical problem of
assessing an unknown distribution. This leads to features of the expert
judgement problem in the design context which are substantially different from
those seen, for example, in risk assessment. In particular, the role of experts
in problem structuring and in developing failure mitigation options is much
more prominent, and there is a need to take into account the reliability
potential for future mitigation measures downstream in the system life cycle.
An overview is given of the stakeholders typically involved in large scale
systems engineering design projects, and this is used to argue the need for
methods that expose potential judgemental biases in order to generate analyses
that can be said to provide rational consensus about uncertainties. Finally, a
number of key points are developed with the aim of moving toward a framework
that provides a holistic method for tracking reliability assessment through the
design process.Comment: This paper commented in: [arXiv:0708.0285], [arXiv:0708.0287],
[arXiv:0708.0288]. Rejoinder in [arXiv:0708.0293]. Published at
http://dx.doi.org/10.1214/088342306000000510 in the Statistical Science
(http://www.imstat.org/sts/) by the Institute of Mathematical Statistics
(http://www.imstat.org
Multistage Accelerated Reliability Growth Testing Model and Data Analysis
Accelerated reliability growth testing has recently received a renewed interest in reliability engineering. The concepts of accelerated testing and reliability growth individually have been used in a variety of applications, either for hardware systems or software systems. The advantage of using a combined strategy is that it could shorten the testing time while maximizing the reliability. In the literature, there are many references related to optimal test design for reliability from either a component level or a system level. In this research, we suggest an approach which conducts accelerated testing at the component level while supporting estimates of reliability at the system level. Our approach helps one decide where and at what level to conduct accelerated test during the system design and testing process. Our approach is designed to reduce testing cost while still demonstrating that system level requirements are met. We do this testing at lower levels in an accelerated environment, where costs are lower, and minimize the amount of testing at the higher integrated system level where it tends to be more expensive
Recommended from our members
An Assessment of PIER Electric Grid Research 2003-2014 White Paper
This white paper describes the circumstances in California around the turn of the 21st century that led the California Energy Commission (CEC) to direct additional Public Interest Energy Research funds to address critical electric grid issues, especially those arising from integrating high penetrations of variable renewable generation with the electric grid. It contains an assessment of the beneficial science and technology advances of the resultant portfolio of electric grid research projects administered under the direction of the CEC by a competitively selected contractor, the University of California’s California Institute for Energy and the Environment, from 2003-2014
Continuous maintenance and the future – Foundations and technological challenges
High value and long life products require continuous maintenance throughout their life cycle to achieve required performance with optimum through-life cost. This paper presents foundations and technologies required to offer the maintenance service. Component and system level degradation science, assessment and modelling along with life cycle ‘big data’ analytics are the two most important knowledge and skill base required for the continuous maintenance. Advanced computing and visualisation technologies will improve efficiency of the maintenance and reduce through-life cost of the product. Future of continuous maintenance within the Industry 4.0 context also identifies the role of IoT, standards and cyber security
Aeronautical Engineering: A special bibliography with indexes, supplement 72, July 1976
This bibliography lists 184 reports, articles, and other documents introduced into the NASA scientific and technical information system in June 1976
Recommended from our members
Prognostics and health management of light emitting diodes
Prognostics is an engineering process of diagnosing, predicting the remaining useful life and estimating the reliability of systems and products. Prognostics and Health Management (PHM) has emerged in the last decade as one of the most efficient approaches in failure prevention, reliability estimation and remaining useful life predictions of various engineering systems and products. Light Emitting Diodes (LEDs) are optoelectronic micro-devices that are now replacing traditional incandescent and fluorescent lighting, as they have many advantages including higher reliability, greater energy efficiency, long life time and faster switching speed. Even though LEDs have high reliability and long life time, manufacturers and lighting systems designers still need to assess the reliability of LED lighting systems and the failures in the LED.
This research provides both experimental and theoretical results that demonstrate the use of prognostics and health monitoring techniques for high power LEDs subjected to harsh operating conditions. Data driven, model driven and fusion prognostics approaches are developed to monitor and identify LED failures, based on the requirement for the light output power. The approaches adopted in this work are validated and can be used to assess the life of an LED lighting system after their deployment based on the power of the light output emitted. The data driven techniques are only based on monitoring selected operational and performance indicators using sensors whereas the model driven technique is based on sensor data as well as on a developed empirical model. Fusion approach is also developed using the data driven and the model driven approaches to the LED. Real-time implementation of developed approaches are also investigated and discussed
- …