186 research outputs found

    Determining Principal Component Cardinality through the Principle of Minimum Description Length

    Full text link
    PCA (Principal Component Analysis) and its variants areubiquitous techniques for matrix dimension reduction and reduced-dimensionlatent-factor extraction. One significant challenge in using PCA, is thechoice of the number of principal components. The information-theoreticMDL (Minimum Description Length) principle gives objective compression-based criteria for model selection, but it is difficult to analytically applyits modern definition - NML (Normalized Maximum Likelihood) - to theproblem of PCA. This work shows a general reduction of NML prob-lems to lower-dimension problems. Applying this reduction, it boundsthe NML of PCA, by terms of the NML of linear regression, which areknown.Comment: LOD 201

    Matching Models Across Abstraction Levels with Gaussian Processes

    Get PDF
    Biological systems are often modelled at different levels of abstraction depending on the particular aims/resources of a study. Such different models often provide qualitatively concordant predictions over specific parametrisations, but it is generally unclear whether model predictions are quantitatively in agreement, and whether such agreement holds for different parametrisations. Here we present a generally applicable statistical machine learning methodology to automatically reconcile the predictions of different models across abstraction levels. Our approach is based on defining a correction map, a random function which modifies the output of a model in order to match the statistics of the output of a different model of the same system. We use two biological examples to give a proof-of-principle demonstration of the methodology, and discuss its advantages and potential further applications.Comment: LNCS forma

    Testing in the incremental design and development of complex products

    Get PDF
    Testing is an important aspect of design and development which consumes significant time and resource in many companies. However, it has received less research attention than many other activities in product development, and especially, very few publications report empirical studies of engineering testing. Such studies are needed to establish the importance of testing and inform the development of pragmatic support methods. This paper combines insights from literature study with findings from three empirical studies of testing. The case studies concern incrementally developed complex products in the automotive domain. A description of testing practice as observed in these studies is provided, confirming that testing activities are used for multiple purposes depending on the context, and are intertwined with design from start to finish of the development process, not done after it as many models depict. Descriptive process models are developed to indicate some of the key insights, and opportunities for further research are suggested

    Large Scale Structure of the Universe

    Full text link
    Galaxies are not uniformly distributed in space. On large scales the Universe displays coherent structure, with galaxies residing in groups and clusters on scales of ~1-3 Mpc/h, which lie at the intersections of long filaments of galaxies that are >10 Mpc/h in length. Vast regions of relatively empty space, known as voids, contain very few galaxies and span the volume in between these structures. This observed large scale structure depends both on cosmological parameters and on the formation and evolution of galaxies. Using the two-point correlation function, one can trace the dependence of large scale structure on galaxy properties such as luminosity, color, stellar mass, and track its evolution with redshift. Comparison of the observed galaxy clustering signatures with dark matter simulations allows one to model and understand the clustering of galaxies and their formation and evolution within their parent dark matter halos. Clustering measurements can determine the parent dark matter halo mass of a given galaxy population, connect observed galaxy populations at different epochs, and constrain cosmological parameters and galaxy evolution models. This chapter describes the methods used to measure the two-point correlation function in both redshift and real space, presents the current results of how the clustering amplitude depends on various galaxy properties, and discusses quantitative measurements of the structures of voids and filaments. The interpretation of these results with current theoretical models is also presented.Comment: Invited contribution to be published in Vol. 8 of book "Planets, Stars, and Stellar Systems", Springer, series editor T. D. Oswalt, volume editor W. C. Keel, v2 includes additional references, updated to match published versio

    Type Ia Supernovae as Stellar Endpoints and Cosmological Tools

    Full text link
    Empirically, Type Ia supernovae are the most useful, precise, and mature tools for determining astronomical distances. Acting as calibrated candles they revealed the presence of dark energy and are being used to measure its properties. However, the nature of the SN Ia explosion, and the progenitors involved, have remained elusive, even after seven decades of research. But now new large surveys are bringing about a paradigm shift --- we can finally compare samples of hundreds of supernovae to isolate critical variables. As a result of this, and advances in modeling, breakthroughs in understanding all aspects of SNe Ia are finally starting to happen.Comment: Invited review for Nature Communications. Final published version. Shortened, update

    "Hook"-calibration of GeneChip-microarrays: Theory and algorithm

    Get PDF
    Abstract Background: The improvement of microarray calibration methods is an essential prerequisite for quantitative expression analysis. This issue requires the formulation of an appropriate model describing the basic relationship between the probe intensity and the specific transcript concentration in a complex environment of competing interactions, the estimation of the magnitude these effects and their correction using the intensity information of a given chip and, finally the development of practicable algorithms which judge the quality of a particular hybridization and estimate the expression degree from the intensity values. Results: We present the so-called hook-calibration method which co-processes the log-difference (delta) and -sum (sigma) of the perfect match (PM) and mismatch (MM) probe-intensities. The MM probes are utilized as an internal reference which is subjected to the same hybridization law as the PM, however with modified characteristics. After sequence-specific affinity correction the method fits the Langmuir-adsorption model to the smoothed delta-versus-sigma plot. The geometrical dimensions of this so-called hook-curve characterize the particular hybridization in terms of simple geometric parameters which provide information about the mean non-specific background intensity, the saturation value, the mean PM/MM-sensitivity gain and the fraction of absent probes. This graphical summary spans a metrics system for expression estimates in natural units such as the mean binding constants and the occupancy of the probe spots. The method is single-chip based, i.e. it separately uses the intensities for each selected chip. Conclusion: The hook-method corrects the raw intensities for the non-specific background hybridization in a sequence-specific manner, for the potential saturation of the probe-spots with bound transcripts and for the sequence-specific binding of specific transcripts. The obtained chip characteristics in combination with the sensitivity corrected probe-intensity values provide expression estimates scaled in natural units which are given by the binding constants of the particular hybridization.</p

    New Constraints (and Motivations) for Abelian Gauge Bosons in the MeV-TeV Mass Range

    Full text link
    We survey the phenomenological constraints on abelian gauge bosons having masses in the MeV to multi-GeV mass range (using precision electroweak measurements, neutrino-electron and neutrino-nucleon scattering, electron and muon anomalous magnetic moments, upsilon decay, beam dump experiments, atomic parity violation, low-energy neutron scattering and primordial nucleosynthesis). We compute their implications for the three parameters that in general describe the low-energy properties of such bosons: their mass and their two possible types of dimensionless couplings (direct couplings to ordinary fermions and kinetic mixing with Standard Model hypercharge). We argue that gauge bosons with very small couplings to ordinary fermions in this mass range are natural in string compactifications and are likely to be generic in theories for which the gravity scale is systematically smaller than the Planck mass - such as in extra-dimensional models - because of the necessity to suppress proton decay. Furthermore, because its couplings are weak, in the low-energy theory relevant to experiments at and below TeV scales the charge gauged by the new boson can appear to be broken, both by classical effects and by anomalies. In particular, if the new gauge charge appears to be anomalous, anomaly cancellation does not also require the introduction of new light fermions in the low-energy theory. Furthermore, the charge can appear to be conserved in the low-energy theory, despite the corresponding gauge boson having a mass. Our results reduce to those of other authors in the special cases where there is no kinetic mixing or there is no direct coupling to ordinary fermions, such as for recently proposed dark-matter scenarios.Comment: 49 pages + appendix, 21 figures. This is the final version which appears in JHE

    The Formation and Evolution of the First Massive Black Holes

    Full text link
    The first massive astrophysical black holes likely formed at high redshifts (z>10) at the centers of low mass (~10^6 Msun) dark matter concentrations. These black holes grow by mergers and gas accretion, evolve into the population of bright quasars observed at lower redshifts, and eventually leave the supermassive black hole remnants that are ubiquitous at the centers of galaxies in the nearby universe. The astrophysical processes responsible for the formation of the earliest seed black holes are poorly understood. The purpose of this review is threefold: (1) to describe theoretical expectations for the formation and growth of the earliest black holes within the general paradigm of hierarchical cold dark matter cosmologies, (2) to summarize several relevant recent observations that have implications for the formation of the earliest black holes, and (3) to look into the future and assess the power of forthcoming observations to probe the physics of the first active galactic nuclei.Comment: 39 pages, review for "Supermassive Black Holes in the Distant Universe", Ed. A. J. Barger, Kluwer Academic Publisher

    The Formation of the First Massive Black Holes

    Full text link
    Supermassive black holes (SMBHs) are common in local galactic nuclei, and SMBHs as massive as several billion solar masses already exist at redshift z=6. These earliest SMBHs may grow by the combination of radiation-pressure-limited accretion and mergers of stellar-mass seed BHs, left behind by the first generation of metal-free stars, or may be formed by more rapid direct collapse of gas in rare special environments where dense gas can accumulate without first fragmenting into stars. This chapter offers a review of these two competing scenarios, as well as some more exotic alternative ideas. It also briefly discusses how the different models may be distinguished in the future by observations with JWST, (e)LISA and other instruments.Comment: 47 pages with 306 references; this review is a chapter in "The First Galaxies - Theoretical Predictions and Observational Clues", Springer Astrophysics and Space Science Library, Eds. T. Wiklind, V. Bromm & B. Mobasher, in pres

    Radiotherapy to the prostate for men with metastatic prostate cancer in the UK and Switzerland: Long-term results from the STAMPEDE randomised controlled trial.

    Get PDF
    BACKGROUND: STAMPEDE has previously reported that radiotherapy (RT) to the prostate improved overall survival (OS) for patients with newly diagnosed prostate cancer with low metastatic burden, but not those with high-burden disease. In this final analysis, we report long-term findings on the primary outcome measure of OS and on the secondary outcome measures of symptomatic local events, RT toxicity events, and quality of life (QoL). METHODS AND FINDINGS: Patients were randomised at secondary care sites in the United Kingdom and Switzerland between January 2013 and September 2016, with 1:1 stratified allocation: 1,029 to standard of care (SOC) and 1,032 to SOC+RT. No masking of the treatment allocation was employed. A total of 1,939 had metastatic burden classifiable, with 42% low burden and 58% high burden, balanced by treatment allocation. Intention-to-treat (ITT) analyses used Cox regression and flexible parametric models (FPMs), adjusted for stratification factors age, nodal involvement, the World Health Organization (WHO) performance status, regular aspirin or nonsteroidal anti-inflammatory drug (NSAID) use, and planned docetaxel use. QoL in the first 2 years on trial was assessed using prospectively collected patient responses to QLQ-30 questionnaire. Patients were followed for a median of 61.3 months. Prostate RT improved OS in patients with low, but not high, metastatic burden (respectively: 202 deaths in SOC versus 156 in SOC+RT, hazard ratio (HR) = 0·64, 95% CI 0.52, 0.79, p < 0.001; 375 SOC versus 386 SOC+RT, HR = 1.11, 95% CI 0.96, 1.28, p = 0·164; interaction p < 0.001). No evidence of difference in time to symptomatic local events was found. There was no evidence of difference in Global QoL or QLQ-30 Summary Score. Long-term urinary toxicity of grade 3 or worse was reported for 10 SOC and 10 SOC+RT; long-term bowel toxicity of grade 3 or worse was reported for 15 and 11, respectively. CONCLUSIONS: Prostate RT improves OS, without detriment in QoL, in men with low-burden, newly diagnosed, metastatic prostate cancer, indicating that it should be recommended as a SOC. TRIAL REGISTRATION: ClinicalTrials.gov NCT00268476, ISRCTN.com ISRCTN78818544
    corecore