92 research outputs found

    Accounting for Source Uncertainties in Analyses of Astronomical Survey Data

    Get PDF
    I discuss an issue arising in analyzing data from astronomical surveys: accounting for measurement uncertainties in the properties of individual sources detected in a survey when making inferences about the entire population of sources. Source uncertainties require the analyst to introduce unknown ``incidental'' parameters for each source. The number of parameters thus grows with the size of the sample, and standard theorems guaranteeing asymptotic convergence of maximum likelihood estimates fail in such settings. From the Bayesian point of view, the missing ingredient in such analyses is accounting for the volume in the incidental parameter space via marginalization. I use simple simulations, motivated by modeling the distribution of trans-Neptunian objects surveyed in the outer solar system, to study the effects of source uncertainties on inferences. The simulations show that current non-Bayesian methods for handling source uncertainties (ignoring them, or using an ad hoc incidental parameter integration) produce incorrect inferences, with errors that grow more severe with increasing sample size. In contrast, accounting for source uncertainty via marginalization leads to sound inferences for any sample size.Comment: 12 pages, 5 figures; to appear in Bayesian Inference And Maximum Entropy Methods In Science And Engineering: 24th International Workshop, Garching, Germany, 2004; ed. Volker Dose et al. (AIP Conference Proceedings Series

    Sines, steps and droplets: Semiparametric Bayesian modeling of arrival time series

    Full text link
    I describe ongoing work developing Bayesian methods for flexible modeling of arrival time series data without binning, aiming to improve detection and measurement of X-ray and gamma-ray pulsars, and of pulses in gamma-ray bursts. The methods use parametric and semiparametric Poisson point process models for the event rate, and by design have close connections to conventional frequentist methods currently used in time-domain astronomy.Comment: 4 pages, 1 figure; to appear in the proceedings of IAU Symposium 285, "New Horizons in Time Domain Astronomy" (proceedings eds. Elizabeth Griffin, Bob Hanisch, and Rob Seaman), Cambridge University Press; see http://www.physics.ox.ac.uk/IAUS285

    Bayesian Adaptive Exploration

    Full text link
    I describe a framework for adaptive scientific exploration based on iterating an Observation--Inference--Design cycle that allows adjustment of hypotheses and observing protocols in response to the results of observation on-the-fly, as data are gathered. The framework uses a unified Bayesian methodology for the inference and design stages: Bayesian inference to quantify what we have learned from the available data and predict future data, and Bayesian decision theory to identify which new observations would teach us the most. When the goal of the experiment is simply to make inferences, the framework identifies a computationally efficient iterative ``maximum entropy sampling'' strategy as the optimal strategy in settings where the noise statistics are independent of signal properties. Results of applying the method to two ``toy'' problems with simulated data--measuring the orbit of an extrasolar planet, and locating a hidden one-dimensional object--show the approach can significantly improve observational efficiency in settings that have well-defined nonlinear models. I conclude with a list of open issues that must be addressed to make Bayesian adaptive exploration a practical and reliable tool for optimizing scientific exploration.Comment: 17 pages, 5 figure

    Introduction to papers on astrostatistics

    Full text link
    We are pleased to present a Special Section on Statistics and Astronomy in this issue of the The Annals of Applied Statistics. Astronomy is an observational rather than experimental science; as a result, astronomical data sets both small and large present particularly challenging problems to analysts who must make the best of whatever the sky offers their instruments. The resulting statistical problems have enormous diversity. In one problem, one may have to carefully quantify uncertainty in a hard-won, sparse data set; in another, the sheer volume of data may forbid a formally optimal analysis, requiring judicious balancing of model sophistication, approximations, and clever algorithms. Often the data bear a complex relationship to the underlying phenomenon producing them, much in the manner of inverse problems.Comment: Published in at http://dx.doi.org/10.1214/09-AOAS234 the Annals of Applied Statistics (http://www.imstat.org/aoas/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Pencil-Beam Surveys for Faint Trans-Neptunian Objects

    Get PDF
    We have conducted pencil-beam searches for outer solar system objects to a limiting magnitude of R ~ 26. Five new trans-neptunian objects were detected in these searches. Our combined data set provides an estimate of ~90 trans-neptunian objects per square degree brighter than ~ 25.9. This estimate is a factor of 3 above the expected number of objects based on an extrapolation of previous surveys with brighter limits, and appears consistent with the hypothesis of a single power-law luminosity function for the entire trans-neptunian region. Maximum likelihood fits to all self-consistent published surveys with published efficiency functions predicts a cumulative sky density Sigma(<R) obeying log10(Sigma) = 0.76(R-23.4) objects per square degree brighter than a given magnitude R.Comment: Accepted by AJ, 18 pages, including 6 figure

    Bayesian Methods for Analysis and Adaptive Scheduling of Exoplanet Observations

    Full text link
    We describe work in progress by a collaboration of astronomers and statisticians developing a suite of Bayesian data analysis tools for extrasolar planet (exoplanet) detection, planetary orbit estimation, and adaptive scheduling of observations. Our work addresses analysis of stellar reflex motion data, where a planet is detected by observing the "wobble" of its host star as it responds to the gravitational tug of the orbiting planet. Newtonian mechanics specifies an analytical model for the resulting time series, but it is strongly nonlinear, yielding complex, multimodal likelihood functions; it is even more complex when multiple planets are present. The parameter spaces range in size from few-dimensional to dozens of dimensions, depending on the number of planets in the system, and the type of motion measured (line-of-sight velocity, or position on the sky). Since orbits are periodic, Bayesian generalizations of periodogram methods facilitate the analysis. This relies on the model being linearly separable, enabling partial analytical marginalization, reducing the dimension of the parameter space. Subsequent analysis uses adaptive Markov chain Monte Carlo methods and adaptive importance sampling to perform the integrals required for both inference (planet detection and orbit measurement), and information-maximizing sequential design (for adaptive scheduling of observations). We present an overview of our current techniques and highlight directions being explored by ongoing research.Comment: 29 pages, 11 figures. An abridged version is accepted for publication in Statistical Methodology for a special issue on astrostatistics, with selected (refereed) papers presented at the Astronomical Data Analysis Conference (ADA VI) held in Monastir, Tunisia, in May 2010. Update corrects equation (3
    • …
    corecore