17,907 research outputs found

    A Quantile Variant of the EM Algorithm and Its Applications to Parameter Estimation with Interval Data

    Full text link
    The expectation-maximization (EM) algorithm is a powerful computational technique for finding the maximum likelihood estimates for parametric models when the data are not fully observed. The EM is best suited for situations where the expectation in each E-step and the maximization in each M-step are straightforward. A difficulty with the implementation of the EM algorithm is that each E-step requires the integration of the log-likelihood function in closed form. The explicit integration can be avoided by using what is known as the Monte Carlo EM (MCEM) algorithm. The MCEM uses a random sample to estimate the integral at each E-step. However, the problem with the MCEM is that it often converges to the integral quite slowly and the convergence behavior can also be unstable, which causes a computational burden. In this paper, we propose what we refer to as the quantile variant of the EM (QEM) algorithm. We prove that the proposed QEM method has an accuracy of O(1/K2)O(1/K^2) while the MCEM method has an accuracy of Op(1/K)O_p(1/\sqrt{K}). Thus, the proposed QEM method possesses faster and more stable convergence properties when compared with the MCEM algorithm. The improved performance is illustrated through the numerical studies. Several practical examples illustrating its use in interval-censored data problems are also provided

    Calculation of Weibull strength parameters and Batdorf flow-density constants for volume- and surface-flaw-induced fracture in ceramics

    Get PDF
    The calculation of shape and scale parameters of the two-parameter Weibull distribution is described using the least-squares analysis and maximum likelihood methods for volume- and surface-flaw-induced fracture in ceramics with complete and censored samples. Detailed procedures are given for evaluating 90 percent confidence intervals for maximum likelihood estimates of shape and scale parameters, the unbiased estimates of the shape parameters, and the Weibull mean values and corresponding standard deviations. Furthermore, the necessary steps are described for detecting outliers and for calculating the Kolmogorov-Smirnov and the Anderson-Darling goodness-of-fit statistics and 90 percent confidence bands about the Weibull distribution. It also shows how to calculate the Batdorf flaw-density constants by uing the Weibull distribution statistical parameters. The techniques described were verified with several example problems, from the open literature, and were coded. The techniques described were verified with several example problems from the open literature, and were coded in the Structural Ceramics Analysis and Reliability Evaluation (SCARE) design program

    Estimating the reliability of composite scores

    Get PDF
    In situations where multiple tests are administered (such as the GCSE subjects), scores from individual tests are frequently combined to produce a composite score. As part of the Ofqual reliability programme, this study, through a review of literature, attempts to: look at the different approaches that are employed to form composite scores from component or unit scores; investigate the implications of the use of the different approaches for the psychometric properties, particularly the reliability, of the composite scores; and identify procedures that are commonly used to estimate the reliability measure of composite scores

    Maximum likelihood estimation in a mixture regression model using the EM algorithm

    Get PDF
    To an extremely difficult problem of finding the maximum likelihood estimates in a specific mixture regression model, a combination of several optimization techniques is found to be useful. These algorithms are the continuation method, Newton-Raphson method, and simplex method. The simplex method finds a globally approximate solution, then a combination of the continuation method and the Newton-Raphson method finds a more accurate solution. In this paper, this combination method is applied to find the maximum likelihood estimates in a Weibull-power-law type regression model, as well as the well-known methods like the EM algorithm, is discussed in this paper

    Software Reliability Models

    Get PDF
    The problem considered here is the building of Non-homogeneous Poisson Process (NHPP) model. Currently existing popular NHPP process models like Goel-Okumoto (G-O) and Yamada et al models suffer from the drawback that the probability density function of the inter-failure times is an improper density function. This is because the event no failure in (0, oo] is allowed in these models. In real life situations we cannot draw sample(s) from such a population and also none of the moments of inter-failure times exist. Therefore, these models are unsuitable for modelling real software error data. On the other hand if the density function of the inter-failure times is made proper by multiplying with a constant, then we cannot assume finite number of expected faults in the system which is the basic assumption in building the software reliability models. Taking these factors into consideration, we have introduced an extra parameter, say c, in both the G -0 and Yamada et al models in order to get a new model. We find that a specific value of this new parameter gives rise to a proper density for inter-failure times. The G -0 and Yamada et al models are special cases of these models corresponding to c = 0. This raises the question - “Can we do better than existing G -0 and Yamada et al models when 0 \u3c c \u3c 1 ?”. The answer is ‘yes’. With this objective, the behavior of the software failure counting process { N ( t ) , t \u3e 0} has been studied. Several measures, such as the number of failures by some prespecified time, the number of errors remaining in the system at a future time, distribution of remaining number of faults in the system and reliability during a mission have been proposed in this research. Maximum likelihood estimation method was used to estimate the parameters. Sufficient conditions for the existence of roots of the ML equations were derived. Some of the important statistical aspects of G -0 and Yamada et al models, like conditions for the existence and uniqueness of the ML equations, were not worked out so far in the literature. We have derived these conditions and proved uniqueness of the roots for these models. Finally four different sets of actual failure time data were analyzed. i

    Models and metrics for software management and engineering

    Get PDF
    This paper attempts to characterize and present a state of the art view of several quantitative models and metrics of the software life cycle. These models and metrics can be used to aid in managing and engineering software projects. They deal with various aspects of the software process and product, including resources allocation and estimation, changes and errors, size, complexity and reliability. Some indication is given of the extent to which the various models have been used and the success they have achieved

    A general introduction to software reliability

    Get PDF
    • 

    corecore