1,313 research outputs found

    A log-Birnbaum-Saunders Regression Model with Asymmetric Errors

    Full text link
    The paper by Leiva et al. (2010) introduced a skewed version of the sinh-normal distribution, discussed some of its properties and characterized an extension of the Birnbaum-Saunders distribution associated with this distribution. In this paper, we introduce a skewed log-Birnbaum-Saunders regression model based on the skewed sinh-normal distribution. Some influence methods, such as the local influence and generalized leverage are presented. Additionally, we derived the normal curvatures of local influence under some perturbation schemes. An empirical application to a real data set is presented in order to illustrate the usefulness of the proposed model.Comment: Submitted for publicatio

    A neural network approach to audio-assisted movie dialogue detection

    Get PDF
    A novel framework for audio-assisted dialogue detection based on indicator functions and neural networks is investigated. An indicator function defines that an actor is present at a particular time instant. The cross-correlation function of a pair of indicator functions and the magnitude of the corresponding cross-power spectral density are fed as input to neural networks for dialogue detection. Several types of artificial neural networks, including multilayer perceptrons, voted perceptrons, radial basis function networks, support vector machines, and particle swarm optimization-based multilayer perceptrons are tested. Experiments are carried out to validate the feasibility of the aforementioned approach by using ground-truth indicator functions determined by human observers on 6 different movies. A total of 41 dialogue instances and another 20 non-dialogue instances is employed. The average detection accuracy achieved is high, ranging between 84.78%±5.499% and 91.43%±4.239%

    STATISTICAL INTERVALS FOR VARIOUS DISTRIBUTIONS BASED ON DIFFERENT INFERENCE METHODS

    Get PDF
    Statistical intervals (e.g., confidence, prediction, or tolerance) are widely used to quantify uncertainty, but complex settings can create challenges to obtain such intervals that possess the desired properties. My thesis will address diverse data settings and approaches that are shown empirically to have good performance. We first introduce a focused treatment on using a single-layer bootstrap calibration to improve the coverage probabilities of two-sided parametric tolerance intervals for non-normal distributions. We then turn to zero-inflated data, which are commonly found in, among other areas, pharmaceutical and quality control applications. However, the inference problem often becomes difficult in the presence of excess zeros. When data are semicontinuous, the log-normal and gamma distributions are often considered for modeling the positive part of the model. The problems of constructing a confidence interval for the mean and calculating an upper tolerance limit of a zero-inflated gamma population are considered using generalized fiducial inference. Furthermore, we use generalized fiducial inference on the problem of constructing confidence intervals for the population mean of zero-inflated Poisson distribution. Birnbaum–Saunders distribution is widely used as a failure time distribution in reliability applications to model failure times. Statistical intervals for Birnbaum–Saunders distribution are not well developed. Moreover, we utilize generalized fiducial inference to obtain the upper prediction limit and upper tolerance limit for Birnbaum–Saunders distribution. Simulation studies and real data examples are used to illustrate the effectiveness of the proposed methods

    Comparison of polybrominated diphenyl ethers (PBDEs) and polychlorinated biphenyls (PCBs) in the serum of hypothyroxinemic and euthyroid dogs.

    Get PDF
    ObjectiveTo determine the profile of 14 polybrominated diphenyl ethers (PBDEs) and 23 polychlorinated biphenyls (PCBs) in serum of domestic canines and whether this was predictive of thyroid hormone status.SamplesSerum samples were collected from 51 client-owned dogs visiting the University of California Davis William R. Pritchard Veterinary Medical Teaching Hospital during 2012 to 2016 for routine appointments. Fifteen dogs were diagnosed with hypothyroxinemia while 36 were euthyroid.ProceduresConcentrations of PBDEs and PCBs in canine serum samples were measured by gas chromatography mass spectrometry. Logistic regression analysis was used to determine the association between the presence/absence of canine hypothyroxinemia and the serum concentration of individual PBDE or PCB congeners.ResultsThe median concentrations of total PBDE and PCB congeners in the hypothyroxinemic group were 660 and 1,371 ng/g lipid, respectively, which were higher than concentrations detected in the control group. However, logistic regression analysis determined that current concentrations of PBDEs and PCBs in canines were not significantly associated with hypothyroxinemia. BDE 183 was the only congener showing near significance (p = 0.068).ConclusionsPBDE and PCB congeners were detected in all canine samples confirming ongoing exposure to these pollutants. Because household dogs share the human environment, they may serve as biosentinels of human exposure to these contaminants

    Mixture of linear experts model for censored data: A novel approach with scale-mixture of normal distributions

    Full text link
    The classical mixture of linear experts (MoE) model is one of the widespread statistical frameworks for modeling, classification, and clustering of data. Built on the normality assumption of the error terms for mathematical and computational convenience, the classical MoE model has two challenges: 1) it is sensitive to atypical observations and outliers, and 2) it might produce misleading inferential results for censored data. The paper is then aimed to resolve these two challenges, simultaneously, by proposing a novel robust MoE model for model-based clustering and discriminant censored data with the scale-mixture of normal class of distributions for the unobserved error terms. Based on this novel model, we develop an analytical expectation-maximization (EM) type algorithm to obtain the maximum likelihood parameter estimates. Simulation studies are carried out to examine the performance, effectiveness, and robustness of the proposed methodology. Finally, real data is used to illustrate the superiority of the new model.Comment: 21 pages

    Statistical Reliability with Applications

    Get PDF
    This chapter reviews fundamental ideas in reliability theory and inference. The first part of the chapter accounts for lifetime distributions that are used in engineering reliability analyis, including general properties of reliability distributions that pertain to lifetime for manufactured products. Certain distributions are formulated on the basis of simple physical properties, and other are more or less empirical. The first part of the chapter ends with a description of graphical and analytical methods to find appropriate lifetime distributions for a set of failure data. The second part of the chapter describes statistical methods for analyzing reliability data, including maximum likelihood estimation and likelihood ratio testing. Degradation data are more prevalent in experiments in which failure is rare and test time is limited. Special regression techniques for degradation data can be used to draw inference on the underlying lifetime distribution, even if failures are rarely observed. The last part of the chapter discusses reliability for systems. Along with the components that comprise the system, reliability analysis must take account of the system configuration and (stochastic) component dependencies. System reliability is illustrated with an analysis of logistics systems (e.g., moving goods in a system of product sources and retail outlets). Robust reliability design can be used to construct a supply chain that runs with maximum efficiency or minimum cost

    Incorporating unobserved heterogeneity and multiple event types in survival models : a Bayesian approach

    Get PDF
    This thesis covers theoretical and practical aspects of Bayesian inference and survival analysis, which is a powerful tool for the analysis of the time until a certain event of interest occurs. This dissertation focuses on non-standard models inspired by features of real datasets that are not accommodated by conventional models. Materials are divided in two parts. The first and more extended part relates to the development of flexible parametric lifetime distributions motivated by the presence of anomalous observations and other forms of unobserved heterogeneity. Chapter 2 presents the use of mixture families of lifetime distributions for this purpose. This idea can be interpreted as the introduction of an observation-specific random effect on the survival distribution. Two families generated via this mechanism are studied in Chapter 3. Covariates are introduced through an accelerated failure times representation, for which the interpretation of the regression coefficients is invariant to the distribution of the random effect. The Bayesian model is completed using reasonable (improper) priors that require a minimum input from practitioners. Under mild conditions, these priors induce a well-defined posterior distribution. In addition, the mixture structure is exploited in order to propose a novel method for outlier detection where anomalous observations are identified via the posterior distribution of the individual-specific random effects. The analysis is illustrated in Chapter 4 using three real medical applications. Chapter 5 comprises the second part of this thesis, which is motivated in the context of university outcomes. The aim of the study is to identify determinants of the length of stay at university and its associated academic outcome for undergraduate students of the Pontificia Universidad Católica de Chile. In this setting, survival times are defined as the time until the end of the enrollment period, which can relate to different reasons - graduation or dropout - that are driven by different processes. Hence, a competing risks model is employed for the analysis. Model uncertainty is handled through Bayesian model averaging, which leads to a better predictive performance than choosing a unique model. The output of this analysis does not account for all features of this complex dataset yet it provides a better understanding of the problem and a starting point for future research. Finally, Chapter 6 summarizes the main findings of this work and suggests future extensions
    corecore