240,427 research outputs found
Recommended from our members
Finite element modelling of atomic force microscope cantilever beams with uncertainty in material and dimensional parameters
Copyright Ā© 2014 by Institute of Fundamental Technological Research
Polish Academy of Sciences, Warsaw, PolandThe stiffness and the natural frequencies of a rectangular and a V-shaped micro-cantilever beams used in Atomic Force Microscope (AFM) were analysed using the Finite Element (FE) method. A determinate analysis in the material and dimensional parameters was first carried out to compare with published analytical and experimental results. Uncertainties in the beamsā parameters such as the material properties and dimensions due to the fabrication process were then modelled using a statistic FE analysis. It is found that for the rectangular micro-beam, a Ā±5% change in the value of the parameters could result in 3 to 8-folds (up to more than 45%) errors in the stiffness or the 1st natural frequency of the cantilever. Such big uncertainties need to be considered in the design and calibration of AFM to ensure the measurement accuracy at the micron and nano scales. In addition, a sensitivity analysis was carried out for the influence of the studied parameters. The finding provides useful guidelines on the design of micro-cantilevers used in the AFM technology.The research was supported by Sichuan International Research Collaboration Project (2014HH0022)
Methodological Issues in Building, Training, and Testing Artificial Neural Networks
We review the use of artificial neural networks, particularly the feedforward
multilayer perceptron with back-propagation for training (MLP), in ecological
modelling. Overtraining on data or giving vague references to how it was
avoided is the major problem. Various methods can be used to determine when to
stop training in artificial neural networks: 1) early stopping based on
cross-validation, 2) stopping after a analyst defined error is reached or after
the error levels off, 3) use of a test data set. We do not recommend the third
method as the test data set is then not independent of model development. Many
studies used the testing data to optimize the model and training. Although this
method may give the best model for that set of data it does not give
generalizability or improve understanding of the study system. The importance
of an independent data set cannot be overemphasized as we found dramatic
differences in model accuracy assessed with prediction accuracy on the training
data set, as estimated with bootstrapping, and from use of an independent data
set. The comparison of the artificial neural network with a general linear
model (GLM) as a standard procedure is recommended because a GLM may perform as
well or better than the MLP. MLP models should not be treated as black box
models but instead techniques such as sensitivity analyses, input variable
relevances, neural interpretation diagrams, randomization tests, and partial
derivatives should be used to make the model more transparent, and further our
ecological understanding which is an important goal of the modelling process.
Based on our experience we discuss how to build a MLP model and how to optimize
the parameters and architecture.Comment: 22 pages, 2 figures. Presented in ISEI3 (2002). Ecological Modelling
in pres
Material parameter estimation and hypothesis testing on a 1D viscoelastic stenosis model: Methodology
This is the post-print version of the final published paper that is available from the link below. Copyright @ 2013 Walter de Gruyter GmbH.Non-invasive detection, localization and characterization of an arterial stenosis (a blockage or partial blockage in the artery) continues to be an important problem in medicine. Partial blockage stenoses are known to generate disturbances in blood flow which generate shear waves in the chest cavity. We examine a one-dimensional viscoelastic model that incorporates KelvināVoigt damping and internal variables, and develop a proof-of-concept methodology using simulated data. We first develop an estimation procedure for the material parameters. We use this procedure to determine confidence intervals for the estimated parameters, which indicates the efficacy of finding parameter estimates in practice. Confidence intervals are computed using asymptotic error theory as well as bootstrapping. We then develop a model comparison test to be used in determining if a particular data set came from a low input amplitude or a high input amplitude; this we anticipate will aid in determining when stenosis is present. These two thrusts together will serve as the methodological basis for our continuing analysis using experimental data currently being collected.National Institute of Allergy and Infectious Diseases, Air Force Office of Scientific Research, Department of Education, and Engineering and Physical Sciences Research Council
The PyCBC search for gravitational waves from compact binary coalescence
We describe the PyCBC search for gravitational waves from compact-object
binary coalescences in advanced gravitational-wave detector data. The search
was used in the first Advanced LIGO observing run and unambiguously identified
two black hole binary mergers, GW150914 and GW151226. At its core, the PyCBC
search performs a matched-filter search for binary merger signals using a bank
of gravitational-wave template waveforms. We provide a complete description of
the search pipeline including the steps used to mitigate the effects of noise
transients in the data, identify candidate events and measure their statistical
significance. The analysis is able to measure false-alarm rates as low as one
per million years, required for confident detection of signals. Using data from
initial LIGO's sixth science run, we show that the new analysis reduces the
background noise in the search, giving a 30% increase in sensitive volume for
binary neutron star systems over previous searches.Comment: 29 pages, 7 figures, accepted by Classical and Quantum Gravit
A non-parametric peak finder algorithm and its application in searches for new physics
We have developed an algorithm for non-parametric fitting and extraction of
statistically significant peaks in the presence of statistical and systematic
uncertainties. Applications of this algorithm for analysis of high-energy
collision data are discussed. In particular, we illustrate how to use this
algorithm in general searches for new physics in invariant-mass spectra using
pp Monte Carlo simulations.Comment: 7 pages, 1 figur
The problem of evaluating automated large-scale evidence aggregators
In the biomedical context, policy makers face a large amount of potentially discordant evidence from different sources. This prompts the question of how this evidence should be aggregated in the interests of best-informed policy recommendations. The starting point of our discussion is Hunter and Williamsā recent work on an automated aggregation method for medical evidence. Our negative claim is that it is far from clear what the relevant criteria for evaluating an evidence aggregator of this sort are. What is the appropriate balance between explicitly coded algorithms and implicit reasoning involved, for instance, in the packaging of input evidence? In short: What is the optimal degree of āautomationā? On the positive side: We propose the ability to perform an adequate robustness analysis as the focal criterion, primarily because it directs efforts to what is most important, namely, the structure of the algorithm and the appropriate extent of automation. Moreover, where there are resource constraints on the aggregation process, one must also consider what balance between volume of evidence and accuracy in the treatment of individual evidence best facilitates inference. There is no prerogative to aggregate the total evidence available if this would in fact reduce overall accuracy
- ā¦