2,359 research outputs found
Benchmarking and parameter sensitivity of physiological and vegetation dynamics using the Functionally Assembled Terrestrial Ecosystem Simulator (FATES) at Barro Colorado Island, Panama
Plant functional traits determine vegetation responses to environmental variation, but variation in trait values is large, even within a single site. Likewise, uncertainty in how these traits map to Earth system feedbacks is large. We use a vegetation demographic model (VDM), the Functionally Assembled Terrestrial Ecosystem Simulator (FATES), to explore parameter sensitivity of model predictions, and comparison to observations, at a tropical forest site: Barro Colorado Island in Panama. We define a single 12-dimensional distribution of plant trait variation, derived primarily from observations in Panama, and define plant functional types (PFTs) as random draws from this distribution. We compare several model ensembles, where individual ensemble members vary only in the plant traits that define PFTs, and separate ensembles differ from each other based on either model structural assumptions or non-trait, ecosystem-level parameters, which include (a) the number of competing PFTs present in any simulation and (b) parameters that govern disturbance and height-based light competition. While single-PFT simulations are roughly consistent with observations of productivity at Barro Colorado Island, increasing the number of competing PFTs strongly shifts model predictions towards higher productivity and biomass forests. Different ecosystem variables show greater sensitivity than others to the number of competing PFTs, with the predictions that are most dominated by large trees, such as biomass, being the most sensitive. Changing disturbance and height-sorting parameters, i.e., the rules of competitive trait filtering, shifts regimes of dominance or coexistence between early- and late-successional PFTs in the model. Increases to the extent or severity of disturbance, or to the degree of determinism in height-based light competition, all act to shift the community towards early-successional PFTs. In turn, these shifts in competitive outcomes alter predictions of ecosystem states and fluxes, with more early-successional-dominated forests having lower biomass. It is thus crucial to differentiate between plant traits, which are under competitive pressure in VDMs, from those model parameters that are not and to better understand the relationships between these two types of model parameters to quantify sources of uncertainty in VDMs
The fallacy of enrolling only high-risk subjects in cancer prevention trials: Is there a "free lunch"?
BACKGROUND: There is a common belief that most cancer prevention trials should be restricted to high-risk subjects in order to increase statistical power. This strategy is appropriate if the ultimate target population is subjects at the same high-risk. However if the target population is the general population, three assumptions may underlie the decision to enroll high-risk subject instead of average-risk subjects from the general population: higher statistical power for the same sample size, lower costs for the same power and type I error, and a correct ratio of benefits to harms. We critically investigate the plausibility of these assumptions. METHODS: We considered each assumption in the context of a simple example. We investigated statistical power for fixed sample size when the investigators assume that relative risk is invariant over risk group, but when, in reality, risk difference is invariant over risk groups. We investigated possible costs when a trial of high-risk subjects has the same power and type I error as a larger trial of average-risk subjects from the general population. We investigated the ratios of benefit to harms when extrapolating from high-risk to average-risk subjects. RESULTS: Appearances here are misleading. First, the increase in statistical power with a trial of high-risk subjects rather than the same number of average-risk subjects from the general population assumes that the relative risk is the same for high-risk and average-risk subjects. However, if the absolute risk difference rather than the relative risk were the same, the power can be less with the high-risk subjects. In the analysis of data from a cancer prevention trial, we found that invariance of absolute risk difference over risk groups was nearly as plausible as invariance of relative risk over risk groups. Therefore a priori assumptions of constant relative risk across risk groups are not robust, limiting extrapolation of estimates of benefit to the general population. Second, a trial of high-risk subjects may cost more than a larger trial of average risk subjects with the same power and type I error because of additional recruitment and diagnostic testing to identify high-risk subjects. Third, the ratio of benefits to harms may be more favorable in high-risk persons than in average-risk persons in the general population, which means that extrapolating this ratio to the general population would be misleading. Thus there is no free lunch when using a trial of high-risk subjects to extrapolate results to the general population. CONCLUSION: Unless the intervention is targeted to only high-risk subjects, cancer prevention trials should be implemented in the general population
Statistical methods in cosmology
The advent of large data-set in cosmology has meant that in the past 10 or 20
years our knowledge and understanding of the Universe has changed not only
quantitatively but also, and most importantly, qualitatively. Cosmologists rely
on data where a host of useful information is enclosed, but is encoded in a
non-trivial way. The challenges in extracting this information must be overcome
to make the most of a large experimental effort. Even after having converged to
a standard cosmological model (the LCDM model) we should keep in mind that this
model is described by 10 or more physical parameters and if we want to study
deviations from it, the number of parameters is even larger. Dealing with such
a high dimensional parameter space and finding parameters constraints is a
challenge on itself. Cosmologists want to be able to compare and combine
different data sets both for testing for possible disagreements (which could
indicate new physics) and for improving parameter determinations. Finally,
cosmologists in many cases want to find out, before actually doing the
experiment, how much one would be able to learn from it. For all these reasons,
sophisiticated statistical techniques are being employed in cosmology, and it
has become crucial to know some statistical background to understand recent
literature in the field. I will introduce some statistical tools that any
cosmologist should know about in order to be able to understand recently
published results from the analysis of cosmological data sets. I will not
present a complete and rigorous introduction to statistics as there are several
good books which are reported in the references. The reader should refer to
those.Comment: 31, pages, 6 figures, notes from 2nd Trans-Regio Winter school in
Passo del Tonale. To appear in Lectures Notes in Physics, "Lectures on
cosmology: Accelerated expansion of the universe" Feb 201
Power laws of complex systems from Extreme physical information
Many complex systems obey allometric, or power, laws y=Yx^{a}. Here y is the
measured value of some system attribute a, Y is a constant, and x is a
stochastic variable. Remarkably, for many living systems the exponent a is
limited to values +or- n/4, n=0,1,2... Here x is the mass of a randomly
selected creature in the population. These quarter-power laws hold for many
attributes, such as pulse rate (n=-1). Allometry has, in the past, been
theoretically justified on a case-by-case basis. An ultimate goal is to find a
common cause for allometry of all types and for both living and nonliving
systems. The principle I - J = extrem. of Extreme physical information (EPI) is
found to provide such a cause. It describes the flow of Fisher information J =>
I from an attribute value a on the cell level to its exterior observation y.
Data y are formed via a system channel function y = f(x,a), with f(x,a) to be
found. Extremizing the difference I - J through variation of f(x,a) results in
a general allometric law f(x,a)= y = Yx^{a}. Darwinian evolution is presumed to
cause a second extremization of I - J, now with respect to the choice of a. The
solution is a=+or-n/4, n=0,1,2..., defining the particular powers of biological
allometry. Under special circumstances, the model predicts that such biological
systems are controlled by but two distinct intracellular information sources.
These sources are conjectured to be cellular DNA and cellular transmembrane ion
gradient
Denitri®cation in a nitrogen-limited stream ecosystem.
Abstract. Denitrification was measured in hyporheic, parafluvial, and bank sediments of Sycamore Creek, Arizona, a nitrogen-limited Sonoran Desert stream. We used three variations of the acetylene block technique to estimate denitrification rates, and compared these estimates to rates of nitrate production through nitrification. Subsurface sediments of Sycamore Creek are typically well-oxygenated, relatively low in nitrate, and low in organic carbon, and therefore are seemingly unlikely sites of denitrification. However, we found that denitrification potential (C & N amended, anaerobic incubations) was substantial, and even by our conservative estimates (unamended, oxic incubations and field chamber nitrous oxide accumulation), denitrification consumed 5-40% of nitrate produced by nitrification. We expected that denitrification would increase along hyporheic and parafluvial flowpaths as dissolved oxygen declined and nitrate increased. To the contrary, we found that denitrification was generally highest at the upstream ends of subsurface flowpaths where surface water had just entered the subsurface zone. This suggests that denitrifiers may be dependent on the import of surface-derived organic matter, resulting in highest denitrification rate at locations of surface-subsurface hydrologic exchange. Laboratory experiments showed that denitrification in Sycamore Creek sediments was primarily nitrogen limited and secondarily carbon limited, and was temperature dependent. Overall, the quantity of nitrate removed from the Sycamore Creek ecosystem via denitrification is significant given the nitrogen-limited status of this stream
Does physical activity counselling enhance the effects of a pedometer-based intervention over the long-term : 12-month findings from the Walking for Wellbeing in the West study
Peer reviewedPublisher PD
Minimax Current Density Coil Design
'Coil design' is an inverse problem in which arrangements of wire are
designed to generate a prescribed magnetic field when energized with electric
current. The design of gradient and shim coils for magnetic resonance imaging
(MRI) are important examples of coil design. The magnetic fields that these
coils generate are usually required to be both strong and accurate. Other
electromagnetic properties of the coils, such as inductance, may be considered
in the design process, which becomes an optimization problem. The maximum
current density is additionally optimized in this work and the resultant coils
are investigated for performance and practicality. Coils with minimax current
density were found to exhibit maximally spread wires and may help disperse
localized regions of Joule heating. They also produce the highest possible
magnetic field strength per unit current for any given surface and wire size.
Three different flavours of boundary element method that employ different basis
functions (triangular elements with uniform current, cylindrical elements with
sinusoidal current and conic section elements with sinusoidal-uniform current)
were used with this approach to illustrate its generality.Comment: 24 pages, 6 figures, 2 tables. To appear in Journal of Physics D:
Applied Physic
Towards Machine Wald
The past century has seen a steady increase in the need of estimating and
predicting complex systems and making (possibly critical) decisions with
limited information. Although computers have made possible the numerical
evaluation of sophisticated statistical models, these models are still designed
\emph{by humans} because there is currently no known recipe or algorithm for
dividing the design of a statistical model into a sequence of arithmetic
operations. Indeed enabling computers to \emph{think} as \emph{humans} have the
ability to do when faced with uncertainty is challenging in several major ways:
(1) Finding optimal statistical models remains to be formulated as a well posed
problem when information on the system of interest is incomplete and comes in
the form of a complex combination of sample data, partial knowledge of
constitutive relations and a limited description of the distribution of input
random variables. (2) The space of admissible scenarios along with the space of
relevant information, assumptions, and/or beliefs, tend to be infinite
dimensional, whereas calculus on a computer is necessarily discrete and finite.
With this purpose, this paper explores the foundations of a rigorous framework
for the scientific computation of optimal statistical estimators/models and
reviews their connections with Decision Theory, Machine Learning, Bayesian
Inference, Stochastic Optimization, Robust Optimization, Optimal Uncertainty
Quantification and Information Based Complexity.Comment: 37 page
The geography of recent genetic ancestry across Europe
The recent genealogical history of human populations is a complex mosaic
formed by individual migration, large-scale population movements, and other
demographic events. Population genomics datasets can provide a window into this
recent history, as rare traces of recent shared genetic ancestry are detectable
due to long segments of shared genomic material. We make use of genomic data
for 2,257 Europeans (the POPRES dataset) to conduct one of the first surveys of
recent genealogical ancestry over the past three thousand years at a
continental scale. We detected 1.9 million shared genomic segments, and used
the lengths of these to infer the distribution of shared ancestors across time
and geography. We find that a pair of modern Europeans living in neighboring
populations share around 10-50 genetic common ancestors from the last 1500
years, and upwards of 500 genetic ancestors from the previous 1000 years. These
numbers drop off exponentially with geographic distance, but since genetic
ancestry is rare, individuals from opposite ends of Europe are still expected
to share millions of common genealogical ancestors over the last 1000 years.
There is substantial regional variation in the number of shared genetic
ancestors: especially high numbers of common ancestors between many eastern
populations likely date to the Slavic and/or Hunnic expansions, while much
lower levels of common ancestry in the Italian and Iberian peninsulas may
indicate weaker demographic effects of Germanic expansions into these areas
and/or more stably structured populations. Recent shared ancestry in modern
Europeans is ubiquitous, and clearly shows the impact of both small-scale
migration and large historical events. Population genomic datasets have
considerable power to uncover recent demographic history, and will allow a much
fuller picture of the close genealogical kinship of individuals across the
world.Comment: Full size figures available from
http://www.eve.ucdavis.edu/~plralph/research.html; or html version at
http://ralphlab.usc.edu/ibd/ibd-paper/ibd-writeup.xhtm
- …