240,427 research outputs found

    Methodological Issues in Building, Training, and Testing Artificial Neural Networks

    Full text link
    We review the use of artificial neural networks, particularly the feedforward multilayer perceptron with back-propagation for training (MLP), in ecological modelling. Overtraining on data or giving vague references to how it was avoided is the major problem. Various methods can be used to determine when to stop training in artificial neural networks: 1) early stopping based on cross-validation, 2) stopping after a analyst defined error is reached or after the error levels off, 3) use of a test data set. We do not recommend the third method as the test data set is then not independent of model development. Many studies used the testing data to optimize the model and training. Although this method may give the best model for that set of data it does not give generalizability or improve understanding of the study system. The importance of an independent data set cannot be overemphasized as we found dramatic differences in model accuracy assessed with prediction accuracy on the training data set, as estimated with bootstrapping, and from use of an independent data set. The comparison of the artificial neural network with a general linear model (GLM) as a standard procedure is recommended because a GLM may perform as well or better than the MLP. MLP models should not be treated as black box models but instead techniques such as sensitivity analyses, input variable relevances, neural interpretation diagrams, randomization tests, and partial derivatives should be used to make the model more transparent, and further our ecological understanding which is an important goal of the modelling process. Based on our experience we discuss how to build a MLP model and how to optimize the parameters and architecture.Comment: 22 pages, 2 figures. Presented in ISEI3 (2002). Ecological Modelling in pres

    Material parameter estimation and hypothesis testing on a 1D viscoelastic stenosis model: Methodology

    Get PDF
    This is the post-print version of the final published paper that is available from the link below. Copyright @ 2013 Walter de Gruyter GmbH.Non-invasive detection, localization and characterization of an arterial stenosis (a blockage or partial blockage in the artery) continues to be an important problem in medicine. Partial blockage stenoses are known to generate disturbances in blood flow which generate shear waves in the chest cavity. We examine a one-dimensional viscoelastic model that incorporates Kelvinā€“Voigt damping and internal variables, and develop a proof-of-concept methodology using simulated data. We first develop an estimation procedure for the material parameters. We use this procedure to determine confidence intervals for the estimated parameters, which indicates the efficacy of finding parameter estimates in practice. Confidence intervals are computed using asymptotic error theory as well as bootstrapping. We then develop a model comparison test to be used in determining if a particular data set came from a low input amplitude or a high input amplitude; this we anticipate will aid in determining when stenosis is present. These two thrusts together will serve as the methodological basis for our continuing analysis using experimental data currently being collected.National Institute of Allergy and Infectious Diseases, Air Force Office of Scientific Research, Department of Education, and Engineering and Physical Sciences Research Council

    The PyCBC search for gravitational waves from compact binary coalescence

    Get PDF
    We describe the PyCBC search for gravitational waves from compact-object binary coalescences in advanced gravitational-wave detector data. The search was used in the first Advanced LIGO observing run and unambiguously identified two black hole binary mergers, GW150914 and GW151226. At its core, the PyCBC search performs a matched-filter search for binary merger signals using a bank of gravitational-wave template waveforms. We provide a complete description of the search pipeline including the steps used to mitigate the effects of noise transients in the data, identify candidate events and measure their statistical significance. The analysis is able to measure false-alarm rates as low as one per million years, required for confident detection of signals. Using data from initial LIGO's sixth science run, we show that the new analysis reduces the background noise in the search, giving a 30% increase in sensitive volume for binary neutron star systems over previous searches.Comment: 29 pages, 7 figures, accepted by Classical and Quantum Gravit

    A non-parametric peak finder algorithm and its application in searches for new physics

    Get PDF
    We have developed an algorithm for non-parametric fitting and extraction of statistically significant peaks in the presence of statistical and systematic uncertainties. Applications of this algorithm for analysis of high-energy collision data are discussed. In particular, we illustrate how to use this algorithm in general searches for new physics in invariant-mass spectra using pp Monte Carlo simulations.Comment: 7 pages, 1 figur

    The problem of evaluating automated large-scale evidence aggregators

    Get PDF
    In the biomedical context, policy makers face a large amount of potentially discordant evidence from different sources. This prompts the question of how this evidence should be aggregated in the interests of best-informed policy recommendations. The starting point of our discussion is Hunter and Williamsā€™ recent work on an automated aggregation method for medical evidence. Our negative claim is that it is far from clear what the relevant criteria for evaluating an evidence aggregator of this sort are. What is the appropriate balance between explicitly coded algorithms and implicit reasoning involved, for instance, in the packaging of input evidence? In short: What is the optimal degree of ā€˜automationā€™? On the positive side: We propose the ability to perform an adequate robustness analysis as the focal criterion, primarily because it directs efforts to what is most important, namely, the structure of the algorithm and the appropriate extent of automation. Moreover, where there are resource constraints on the aggregation process, one must also consider what balance between volume of evidence and accuracy in the treatment of individual evidence best facilitates inference. There is no prerogative to aggregate the total evidence available if this would in fact reduce overall accuracy
    • ā€¦
    corecore