2,359 research outputs found

    Benchmarking and parameter sensitivity of physiological and vegetation dynamics using the Functionally Assembled Terrestrial Ecosystem Simulator (FATES) at Barro Colorado Island, Panama

    Get PDF
    Plant functional traits determine vegetation responses to environmental variation, but variation in trait values is large, even within a single site. Likewise, uncertainty in how these traits map to Earth system feedbacks is large. We use a vegetation demographic model (VDM), the Functionally Assembled Terrestrial Ecosystem Simulator (FATES), to explore parameter sensitivity of model predictions, and comparison to observations, at a tropical forest site: Barro Colorado Island in Panama. We define a single 12-dimensional distribution of plant trait variation, derived primarily from observations in Panama, and define plant functional types (PFTs) as random draws from this distribution. We compare several model ensembles, where individual ensemble members vary only in the plant traits that define PFTs, and separate ensembles differ from each other based on either model structural assumptions or non-trait, ecosystem-level parameters, which include (a) the number of competing PFTs present in any simulation and (b) parameters that govern disturbance and height-based light competition. While single-PFT simulations are roughly consistent with observations of productivity at Barro Colorado Island, increasing the number of competing PFTs strongly shifts model predictions towards higher productivity and biomass forests. Different ecosystem variables show greater sensitivity than others to the number of competing PFTs, with the predictions that are most dominated by large trees, such as biomass, being the most sensitive. Changing disturbance and height-sorting parameters, i.e., the rules of competitive trait filtering, shifts regimes of dominance or coexistence between early- and late-successional PFTs in the model. Increases to the extent or severity of disturbance, or to the degree of determinism in height-based light competition, all act to shift the community towards early-successional PFTs. In turn, these shifts in competitive outcomes alter predictions of ecosystem states and fluxes, with more early-successional-dominated forests having lower biomass. It is thus crucial to differentiate between plant traits, which are under competitive pressure in VDMs, from those model parameters that are not and to better understand the relationships between these two types of model parameters to quantify sources of uncertainty in VDMs

    The fallacy of enrolling only high-risk subjects in cancer prevention trials: Is there a "free lunch"?

    Get PDF
    BACKGROUND: There is a common belief that most cancer prevention trials should be restricted to high-risk subjects in order to increase statistical power. This strategy is appropriate if the ultimate target population is subjects at the same high-risk. However if the target population is the general population, three assumptions may underlie the decision to enroll high-risk subject instead of average-risk subjects from the general population: higher statistical power for the same sample size, lower costs for the same power and type I error, and a correct ratio of benefits to harms. We critically investigate the plausibility of these assumptions. METHODS: We considered each assumption in the context of a simple example. We investigated statistical power for fixed sample size when the investigators assume that relative risk is invariant over risk group, but when, in reality, risk difference is invariant over risk groups. We investigated possible costs when a trial of high-risk subjects has the same power and type I error as a larger trial of average-risk subjects from the general population. We investigated the ratios of benefit to harms when extrapolating from high-risk to average-risk subjects. RESULTS: Appearances here are misleading. First, the increase in statistical power with a trial of high-risk subjects rather than the same number of average-risk subjects from the general population assumes that the relative risk is the same for high-risk and average-risk subjects. However, if the absolute risk difference rather than the relative risk were the same, the power can be less with the high-risk subjects. In the analysis of data from a cancer prevention trial, we found that invariance of absolute risk difference over risk groups was nearly as plausible as invariance of relative risk over risk groups. Therefore a priori assumptions of constant relative risk across risk groups are not robust, limiting extrapolation of estimates of benefit to the general population. Second, a trial of high-risk subjects may cost more than a larger trial of average risk subjects with the same power and type I error because of additional recruitment and diagnostic testing to identify high-risk subjects. Third, the ratio of benefits to harms may be more favorable in high-risk persons than in average-risk persons in the general population, which means that extrapolating this ratio to the general population would be misleading. Thus there is no free lunch when using a trial of high-risk subjects to extrapolate results to the general population. CONCLUSION: Unless the intervention is targeted to only high-risk subjects, cancer prevention trials should be implemented in the general population

    Statistical methods in cosmology

    Full text link
    The advent of large data-set in cosmology has meant that in the past 10 or 20 years our knowledge and understanding of the Universe has changed not only quantitatively but also, and most importantly, qualitatively. Cosmologists rely on data where a host of useful information is enclosed, but is encoded in a non-trivial way. The challenges in extracting this information must be overcome to make the most of a large experimental effort. Even after having converged to a standard cosmological model (the LCDM model) we should keep in mind that this model is described by 10 or more physical parameters and if we want to study deviations from it, the number of parameters is even larger. Dealing with such a high dimensional parameter space and finding parameters constraints is a challenge on itself. Cosmologists want to be able to compare and combine different data sets both for testing for possible disagreements (which could indicate new physics) and for improving parameter determinations. Finally, cosmologists in many cases want to find out, before actually doing the experiment, how much one would be able to learn from it. For all these reasons, sophisiticated statistical techniques are being employed in cosmology, and it has become crucial to know some statistical background to understand recent literature in the field. I will introduce some statistical tools that any cosmologist should know about in order to be able to understand recently published results from the analysis of cosmological data sets. I will not present a complete and rigorous introduction to statistics as there are several good books which are reported in the references. The reader should refer to those.Comment: 31, pages, 6 figures, notes from 2nd Trans-Regio Winter school in Passo del Tonale. To appear in Lectures Notes in Physics, "Lectures on cosmology: Accelerated expansion of the universe" Feb 201

    Power laws of complex systems from Extreme physical information

    Full text link
    Many complex systems obey allometric, or power, laws y=Yx^{a}. Here y is the measured value of some system attribute a, Y is a constant, and x is a stochastic variable. Remarkably, for many living systems the exponent a is limited to values +or- n/4, n=0,1,2... Here x is the mass of a randomly selected creature in the population. These quarter-power laws hold for many attributes, such as pulse rate (n=-1). Allometry has, in the past, been theoretically justified on a case-by-case basis. An ultimate goal is to find a common cause for allometry of all types and for both living and nonliving systems. The principle I - J = extrem. of Extreme physical information (EPI) is found to provide such a cause. It describes the flow of Fisher information J => I from an attribute value a on the cell level to its exterior observation y. Data y are formed via a system channel function y = f(x,a), with f(x,a) to be found. Extremizing the difference I - J through variation of f(x,a) results in a general allometric law f(x,a)= y = Yx^{a}. Darwinian evolution is presumed to cause a second extremization of I - J, now with respect to the choice of a. The solution is a=+or-n/4, n=0,1,2..., defining the particular powers of biological allometry. Under special circumstances, the model predicts that such biological systems are controlled by but two distinct intracellular information sources. These sources are conjectured to be cellular DNA and cellular transmembrane ion gradient

    Denitri®cation in a nitrogen-limited stream ecosystem.

    Get PDF
    Abstract. Denitrification was measured in hyporheic, parafluvial, and bank sediments of Sycamore Creek, Arizona, a nitrogen-limited Sonoran Desert stream. We used three variations of the acetylene block technique to estimate denitrification rates, and compared these estimates to rates of nitrate production through nitrification. Subsurface sediments of Sycamore Creek are typically well-oxygenated, relatively low in nitrate, and low in organic carbon, and therefore are seemingly unlikely sites of denitrification. However, we found that denitrification potential (C & N amended, anaerobic incubations) was substantial, and even by our conservative estimates (unamended, oxic incubations and field chamber nitrous oxide accumulation), denitrification consumed 5-40% of nitrate produced by nitrification. We expected that denitrification would increase along hyporheic and parafluvial flowpaths as dissolved oxygen declined and nitrate increased. To the contrary, we found that denitrification was generally highest at the upstream ends of subsurface flowpaths where surface water had just entered the subsurface zone. This suggests that denitrifiers may be dependent on the import of surface-derived organic matter, resulting in highest denitrification rate at locations of surface-subsurface hydrologic exchange. Laboratory experiments showed that denitrification in Sycamore Creek sediments was primarily nitrogen limited and secondarily carbon limited, and was temperature dependent. Overall, the quantity of nitrate removed from the Sycamore Creek ecosystem via denitrification is significant given the nitrogen-limited status of this stream

    Minimax Current Density Coil Design

    Full text link
    'Coil design' is an inverse problem in which arrangements of wire are designed to generate a prescribed magnetic field when energized with electric current. The design of gradient and shim coils for magnetic resonance imaging (MRI) are important examples of coil design. The magnetic fields that these coils generate are usually required to be both strong and accurate. Other electromagnetic properties of the coils, such as inductance, may be considered in the design process, which becomes an optimization problem. The maximum current density is additionally optimized in this work and the resultant coils are investigated for performance and practicality. Coils with minimax current density were found to exhibit maximally spread wires and may help disperse localized regions of Joule heating. They also produce the highest possible magnetic field strength per unit current for any given surface and wire size. Three different flavours of boundary element method that employ different basis functions (triangular elements with uniform current, cylindrical elements with sinusoidal current and conic section elements with sinusoidal-uniform current) were used with this approach to illustrate its generality.Comment: 24 pages, 6 figures, 2 tables. To appear in Journal of Physics D: Applied Physic

    Towards Machine Wald

    Get PDF
    The past century has seen a steady increase in the need of estimating and predicting complex systems and making (possibly critical) decisions with limited information. Although computers have made possible the numerical evaluation of sophisticated statistical models, these models are still designed \emph{by humans} because there is currently no known recipe or algorithm for dividing the design of a statistical model into a sequence of arithmetic operations. Indeed enabling computers to \emph{think} as \emph{humans} have the ability to do when faced with uncertainty is challenging in several major ways: (1) Finding optimal statistical models remains to be formulated as a well posed problem when information on the system of interest is incomplete and comes in the form of a complex combination of sample data, partial knowledge of constitutive relations and a limited description of the distribution of input random variables. (2) The space of admissible scenarios along with the space of relevant information, assumptions, and/or beliefs, tend to be infinite dimensional, whereas calculus on a computer is necessarily discrete and finite. With this purpose, this paper explores the foundations of a rigorous framework for the scientific computation of optimal statistical estimators/models and reviews their connections with Decision Theory, Machine Learning, Bayesian Inference, Stochastic Optimization, Robust Optimization, Optimal Uncertainty Quantification and Information Based Complexity.Comment: 37 page

    The geography of recent genetic ancestry across Europe

    Get PDF
    The recent genealogical history of human populations is a complex mosaic formed by individual migration, large-scale population movements, and other demographic events. Population genomics datasets can provide a window into this recent history, as rare traces of recent shared genetic ancestry are detectable due to long segments of shared genomic material. We make use of genomic data for 2,257 Europeans (the POPRES dataset) to conduct one of the first surveys of recent genealogical ancestry over the past three thousand years at a continental scale. We detected 1.9 million shared genomic segments, and used the lengths of these to infer the distribution of shared ancestors across time and geography. We find that a pair of modern Europeans living in neighboring populations share around 10-50 genetic common ancestors from the last 1500 years, and upwards of 500 genetic ancestors from the previous 1000 years. These numbers drop off exponentially with geographic distance, but since genetic ancestry is rare, individuals from opposite ends of Europe are still expected to share millions of common genealogical ancestors over the last 1000 years. There is substantial regional variation in the number of shared genetic ancestors: especially high numbers of common ancestors between many eastern populations likely date to the Slavic and/or Hunnic expansions, while much lower levels of common ancestry in the Italian and Iberian peninsulas may indicate weaker demographic effects of Germanic expansions into these areas and/or more stably structured populations. Recent shared ancestry in modern Europeans is ubiquitous, and clearly shows the impact of both small-scale migration and large historical events. Population genomic datasets have considerable power to uncover recent demographic history, and will allow a much fuller picture of the close genealogical kinship of individuals across the world.Comment: Full size figures available from http://www.eve.ucdavis.edu/~plralph/research.html; or html version at http://ralphlab.usc.edu/ibd/ibd-paper/ibd-writeup.xhtm
    corecore