2,838 research outputs found

    Accelerating epistasis analysis in human genetics with consumer graphics hardware

    Get PDF
    BACKGROUND: Human geneticists are now capable of measuring more than one million DNA sequence variations from across the human genome. The new challenge is to develop computationally feasible methods capable of analyzing these data for associations with common human disease, particularly in the context of epistasis. Epistasis describes the situation where multiple genes interact in a complex non-linear manner to determine an individual's disease risk and is thought to be ubiquitous for common diseases. Multifactor Dimensionality Reduction (MDR) is an algorithm capable of detecting epistasis. An exhaustive analysis with MDR is often computationally expensive, particularly for high order interactions. This challenge has previously been met with parallel computation and expensive hardware. The option we examine here exploits commodity hardware designed for computer graphics. In modern computers Graphics Processing Units (GPUs) have more memory bandwidth and computational capability than Central Processing Units (CPUs) and are well suited to this problem. Advances in the video game industry have led to an economy of scale creating a situation where these powerful components are readily available at very low cost. Here we implement and evaluate the performance of the MDR algorithm on GPUs. Of primary interest are the time required for an epistasis analysis and the price to performance ratio of available solutions. FINDINGS: We found that using MDR on GPUs consistently increased performance per machine over both a feature rich Java software package and a C++ cluster implementation. The performance of a GPU workstation running a GPU implementation reduces computation time by a factor of 160 compared to an 8-core workstation running the Java implementation on CPUs. This GPU workstation performs similarly to 150 cores running an optimized C++ implementation on a Beowulf cluster. Furthermore this GPU system provides extremely cost effective performance while leaving the CPU available for other tasks. The GPU workstation containing three GPUs costs 2000whileobtainingsimilarperformanceonaBeowulfclusterrequires150CPUcoreswhich,includingtheaddedinfrastructureandsupportcostoftheclustersystem,costapproximately2000 while obtaining similar performance on a Beowulf cluster requires 150 CPU cores which, including the added infrastructure and support cost of the cluster system, cost approximately 82,500. CONCLUSION: Graphics hardware based computing provides a cost effective means to perform genetic analysis of epistasis using MDR on large datasets without the infrastructure of a computing cluster

    Inner-shelf circulation and sediment dynamics on a series of shoreface-connected ridges offshore of Fire Island, NY

    Get PDF
    Locations along the inner-continental shelf offshore of Fire Island, NY, are characterized by a series of shoreface-connected ridges (SFCRs). These sand ridges have approximate dimensions of 10 km in length, 3 km spacing, and up to similar to 8 m ridge to trough relief and are oriented obliquely at approximately 30 degrees clockwise from the coastline. Stability analysis from previous studies explains how sand ridges such as these could be formed and maintained by storm-driven flows directed alongshore with a key maintenance mechanism of offshore deflected flows over ridge crests and onshore in the troughs. We examine these processes both with a limited set of idealized numerical simulations and analysis of observational data. Model results confirm that alongshore flows over the SFCRs exhibit offshore veering of currents over the ridge crests and onshore-directed flows in the troughs, and demonstrate the opposite circulation pattern for a reverse wind. To further investigate these maintenance processes, oceanographic instruments were deployed at seven sites on the SFCRs offshore of Fire Island to measure water levels, ocean currents, waves, suspended sediment concentrations, and bottom stresses from January to April 2012. Data analysis reveals that during storms with winds from the northeast, the processes of offshore deflection of currents over ridge crests and onshore in the troughs were observed, and during storm events with winds from the southwest, a reverse flow pattern over the ridges occurred. Computations of suspended sediment fluxes identify periods that are consistent with SFCR maintenance mechanisms. Alongshore winds from the northeast drove fluxes offshore on the ridge crest and onshore in the trough that would tend to promote ridge maintenance. However, alongshore winds from the southwest drove opposite circulations. The wind fields are related to different storm types that occur in the region (low-pressure systems, cold fronts, and warm fronts). From the limited data set, we identify that low-pressure systems drive sediment fluxes that tend to promote stability and maintain the SFCRs while cold front type storms appear to drive circulations that are in the opposite sense and may not be a supporting mechanism for ridge maintenance

    A discrete geometric model of concurrent program execution

    Get PDF
    A trace of the execution of a concurrent object-oriented program can be displayed in two-dimensions as a diagram of a non-metric finite geometry. The actions of a programs are represented by points, its objects and threads by vertical lines, its transactions by horizontal lines, its communications and resource sharing by sloping arrows, and its partial traces by rectangular figures. We prove informally that the geometry satisfies the laws of Concurrent Kleene Algebra (CKA); these describe and justify the interleaved implementation of multithreaded programs on computer systems with a lesser number of concurrent processors. More familiar forms of semantics (e.g., verification-oriented and operational) can be derived from CKA. Programs are represented as sets of all their possible traces of execution, and non-determinism is introduced as union of these sets. The geometry is extended to multiple levels of abstraction and granularity; a method call at a higher level can be modelled by a specification of the method body, which is implemented at a lower level. The final section describes how the axioms and definitions of the geometry have been encoded in the interactive proof tool Isabelle, and reports on progress towards automatic checking of the proofs in the paper

    Fluctuation induces evolutionary branching in a modeled microbial ecosystem

    Get PDF
    The impact of environmental fluctuation on species diversity is studied with a model of the evolutionary ecology of microorganisms. We show that environmental fluctuation induces evolutionary branching and assures the consequential coexistence of multiple species. Pairwise invasibility analysis is applied to illustrate the speciation process. We also discuss how fluctuation affects species diversity.Comment: 4 pages, 4 figures. Submitted to Physical Review Letter

    Diabetes status and post-load plasma glucose concentration in relation to site-specific cancer mortality: findings from the original Whitehall study

    Get PDF
    ObjectiveWhile several studies have reported on the relation of diabetes status with pancreatic cancer risk, the predictive value of this disorder for other malignancies is unclear. Methods: The Whitehall study, a 25year follow-up for mortality experience of 18,006 men with data on post-challenge blood glucose and self-reported diabetes, allowed us to address these issues. Results: There were 2158 cancer deaths at follow-up. Of the 15 cancer outcomes, diabetes status was positively associated with mortality from carcinoma of the pancreas and liver, while the relationship with lung cancer was inverse, after controlling for a range of potential covariates and mediators which included obesity and socioeconomic position. After excluding deaths occurring in the first 10years of follow-up to examine the effect of reverse causality, the magnitude of the relationships for carcinoma of the pancreas and lung was little altered, while for liver cancer it was markedly attenuated. Conclusions: In the present study, diabetes status was related to pancreatic, liver, and lung cancer risk. Cohorts with serially collected data on blood glucose and covariates are required to further examine this area

    Modelling the nucleon wave function from soft and hard processes

    Get PDF
    Current light-cone wave functions for the nucleon are unsatisfactory since they are in conflict with the data of the nucleon's Dirac form factor at large momentum transfer. Therefore, we attempt a determination of a new wave function respecting theoretical ideas on its parameterization and satisfying the following constraints: It should provide a soft Feynman contribution to the proton's form factor in agreement with data; it should be consistent with current parameterizations of the valence quark distribution functions and lastly it should provide an acceptable value for the \jp \to N \bar N decay width. The latter process is calculated within the modified perturbative approach to hard exclusive reactions. A simultaneous fit to the three sets of data leads to a wave function whose xx-dependent part, the distribution amplitude, shows the same type of asymmetry as those distribution amplitudes constrained by QCD sum rules. The asymmetry is however much more moderate as in those amplitudes. Our distribution amplitude resembles the asymptotic one in shape but the position of the maximum is somewhat shifted.Comment: 32 pages RevTex + PS-file with 5 figures in uu-encoded, compressed fil

    Traditional Cantonese diet and nasopharyngeal carcinoma risk: a large-scale case-control study in Guangdong, China

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>Nasopharyngeal carcinoma (NPC) is rare in most parts of the world but is a common malignancy in southern China, especially in Guangdong. Dietary habit is regarded as an important modifier of NPC risk in several endemic areas and may partially explain the geographic distribution of NPC incidence. In China, rapid economic development during the past few decades has changed the predominant lifestyle and dietary habits of the Chinese considerably, requiring a reassessment of diet and its potential influence on NPC risk in this NPC-endemic area.</p> <p>Methods</p> <p>To evaluate the association between dietary factors and NPC risk in Guangdong, China, a large-scale, hospital-based case-control study was conducted. 1387 eligible cases and 1459 frequency matched controls were recruited. Odds ratios (ORs) and the corresponding 95% confidence intervals (CIs) were estimated using a logistic regression model, adjusting for age, sex, education, dialect, and habitation household type.</p> <p>Results</p> <p>Observations made include the following: 1) consumption of canton-style salted fish, preserved vegetables and preserved/cured meat were significantly associated with increased risk of NPC, with enhanced odds ratios (OR) of 2.45 (95% CI: 2.03-2.94), 3.17(95% CI: 2.68-3.77) and 2.09 (95% CI: 1.22-3.60) respectively in the highest intake frequency stratum during childhood; 2) consumption of fresh fruit was associated with reduced risk with a dose-dependent relationship (p = 0.001); and 3) consumption of Canton-style herbal tea and herbal slow-cooked soup was associated with decreased risk, with ORs of 0.84 (95% CI: 0.68-1.03) and 0.58 (95% CI: 0.47-0.72) respectively in the highest intake frequency stratum. In multivariate analyses, these associations remained significant.</p> <p>Conclusions</p> <p>It can be inferred that previously established dietary risk factors in the Cantonese population are still stable and have contributed to the incidence of NPC.</p

    Semi-sparse PCA

    Get PDF
    It is well-known that the classical exploratory factor analysis (EFA) of data with more observations than variables has several types of indeterminacy. We study the factor indeterminacy and show some new aspects of this problem by considering EFA as a specific data matrix decomposition. We adopt a new approach to the EFA estimation and achieve a new characterization of the factor indeterminacy problem. A new alternative model is proposed, which gives determinate factors and can be seen as a semi-sparse principal component analysis (PCA). An alternating algorithm is developed, where in each step a Procrustes problem is solved. It is demonstrated that the new model/algorithm can act as a specific sparse PCA and as a low-rank-plus-sparse matrix decomposition. Numerical examples with several large data sets illustrate the versatility of the new model, and the performance and behaviour of its algorithmic implementation

    Precision Measurement of the Newtonian Gravitational Constant Using Cold Atoms

    Full text link
    About 300 experiments have tried to determine the value of the Newtonian gravitational constant, G, so far, but large discrepancies in the results have made it impossible to know its value precisely. The weakness of the gravitational interaction and the impossibility of shielding the effects of gravity make it very difficult to measure G while keeping systematic effects under control. Most previous experiments performed were based on the torsion pendulum or torsion balance scheme as in the experiment by Cavendish in 1798, and in all cases macroscopic masses were used. Here we report the precise determination of G using laser-cooled atoms and quantum interferometry. We obtain the value G=6.67191(99) x 10^(-11) m^3 kg^(-1) s^(-2) with a relative uncertainty of 150 parts per million (the combined standard uncertainty is given in parentheses). Our value differs by 1.5 combined standard deviations from the current recommended value of the Committee on Data for Science and Technology. A conceptually different experiment such as ours helps to identify the systematic errors that have proved elusive in previous experiments, thus improving the confidence in the value of G. There is no definitive relationship between G and the other fundamental constants, and there is no theoretical prediction for its value, against which to test experimental results. Improving the precision with which we know G has not only a pure metrological interest, but is also important because of the key role that G has in theories of gravitation, cosmology, particle physics and astrophysics and in geophysical models.Comment: 3 figures, 1 tabl

    Terahertz underdamped vibrational motion governs protein-ligand binding in solution

    Get PDF
    Low-frequency collective vibrational modes in proteins have been proposed as being responsible for efficiently directing biochemical reactions and biological energy transport. However, evidence of the existence of delocalized vibrational modes is scarce and proof of their involvement in biological function absent. Here we apply extremely sensitive femtosecond optical Kerr-effect spectroscopy to study the depolarized Raman spectra of lysozyme and its complex with the inhibitor triacetylchitotriose in solution. Underdamped delocalized vibrational modes in the terahertz frequency domain are identified and shown to blue-shift and strengthen upon inhibitor binding. This demonstrates that the ligand-binding coordinate in proteins is underdamped and not simply solvent-controlled as previously assumed. The presence of such underdamped delocalized modes in proteins may have significant implications for the understanding of the efficiency of ligand binding and protein–molecule interactions, and has wider implications for biochemical reactivity and biological function
    corecore