234 research outputs found
Epigenetic modification of the oxytocin and glucocorticoid receptor genes is linked to attachment avoidance in young adults
Attachment in the context of intimate pair bonds is most frequently studied in terms of the universal strategy to draw near, or away, from significant others at moments of personal distress. However, important interindividual differences in the quality of attachment exist, usually captured through secure versus insecure – anxious and/or avoidant – attachment orientations. Since Bowlby’s pioneering writings on the theory of attachment, it has been assumed that attachment orientations are influenced by both genetic and social factors – what we would today describe and measure as gene by environment interaction mediated by epigenetic DNA modification – but research in humans on this topic remains extremely limited. We for the first time examined relations between intra-individual differences in attachment and epigenetic modification of the oxytocin receptor (OXTR) and glucocorticoid receptor (NR3C1) gene promoter in 109 young adult human participants. Our results revealed that attachment avoidance was significantly and specifically associated with increased OXTR and NR3C1 promoter methylation. These findings offer first tentative clues on the possible etiology of attachment avoidance in humans by showing epigenetic modification in genes related to both social stress regulation and HPA axis functioning
Learning and generation of long-range correlated sequences
We study the capability to learn and to generate long-range, power-law
correlated sequences by a fully connected asymmetric network. The focus is set
on the ability of neural networks to extract statistical features from a
sequence. We demonstrate that the average power-law behavior is learnable,
namely, the sequence generated by the trained network obeys the same
statistical behavior. The interplay between a correlated weight matrix and the
sequence generated by such a network is explored. A weight matrix with a
power-law correlation function along the vertical direction, gives rise to a
sequence with a similar statistical behavior.Comment: 5 pages, 3 figures, accepted for publication in Physical Review
The dynamics of proving uncolourability of large random graphs I. Symmetric Colouring Heuristic
We study the dynamics of a backtracking procedure capable of proving
uncolourability of graphs, and calculate its average running time T for sparse
random graphs, as a function of the average degree c and the number of vertices
N. The analysis is carried out by mapping the history of the search process
onto an out-of-equilibrium (multi-dimensional) surface growth problem. The
growth exponent of the average running time is quantitatively predicted, in
agreement with simulations.Comment: 5 figure
Multi-Player and Multi-Choice Quantum Game
We investigate a multi-player and multi-choice quantum game. We start from
two-player and two-choice game and the result is better than its classical
version. Then we extend it to N-player and N-choice cases. In the quantum
domain, we provide a strategy with which players can always avoid the worst
outcome. Also, by changing the value of the parameter of the initial state, the
probabilities for players to obtain the best payoff will be much higher that in
its classical version.Comment: 4 pages, 1 figur
Predictive gene lists for breast cancer prognosis: A topographic visualisation study
<p>Abstract</p> <p>Background</p> <p>The controversy surrounding the non-uniqueness of predictive gene lists (PGL) of small selected subsets of genes from very large potential candidates as available in DNA microarray experiments is now widely acknowledged <abbrgrp><abbr bid="B1">1</abbr></abbrgrp>. Many of these studies have focused on constructing discriminative semi-parametric models and as such are also subject to the issue of random correlations of sparse model selection in high dimensional spaces. In this work we outline a different approach based around an unsupervised patient-specific nonlinear topographic projection in predictive gene lists.</p> <p>Methods</p> <p>We construct nonlinear topographic projection maps based on inter-patient gene-list relative dissimilarities. The Neuroscale, the Stochastic Neighbor Embedding(SNE) and the Locally Linear Embedding(LLE) techniques have been used to construct two-dimensional projective visualisation plots of 70 dimensional PGLs per patient, classifiers are also constructed to identify the prognosis indicator of each patient using the resulting projections from those visualisation techniques and investigate whether <it>a-posteriori </it>two prognosis groups are separable on the evidence of the gene lists.</p> <p>A literature-proposed predictive gene list for breast cancer is benchmarked against a separate gene list using the above methods. Generalisation ability is investigated by using the mapping capability of Neuroscale to visualise the follow-up study, but based on the projections derived from the original dataset.</p> <p>Results</p> <p>The results indicate that small subsets of patient-specific PGLs have insufficient prognostic dissimilarity to permit a distinction between two prognosis patients. Uncertainty and diversity across multiple gene expressions prevents unambiguous or even confident patient grouping. Comparative projections across different PGLs provide similar results.</p> <p>Conclusion</p> <p>The random correlation effect to an arbitrary outcome induced by small subset selection from very high dimensional interrelated gene expression profiles leads to an outcome with associated uncertainty. This continuum and uncertainty precludes any attempts at constructing discriminative classifiers.</p> <p>However a patient's gene expression profile could possibly be used in treatment planning, based on knowledge of other patients' responses.</p> <p>We conclude that many of the patients involved in such medical studies are <it>intrinsically unclassifiable </it>on the basis of provided PGL evidence. This additional category of 'unclassifiable' should be accommodated within medical decision support systems if serious errors and unnecessary adjuvant therapy are to be avoided.</p
The influence of feature selection methods on accuracy, stability and interpretability of molecular signatures
Motivation: Biomarker discovery from high-dimensional data is a crucial
problem with enormous applications in biology and medicine. It is also
extremely challenging from a statistical viewpoint, but surprisingly few
studies have investigated the relative strengths and weaknesses of the plethora
of existing feature selection methods. Methods: We compare 32 feature selection
methods on 4 public gene expression datasets for breast cancer prognosis, in
terms of predictive performance, stability and functional interpretability of
the signatures they produce. Results: We observe that the feature selection
method has a significant influence on the accuracy, stability and
interpretability of signatures. Simple filter methods generally outperform more
complex embedded or wrapper methods, and ensemble feature selection has
generally no positive effect. Overall a simple Student's t-test seems to
provide the best results. Availability: Code and data are publicly available at
http://cbio.ensmp.fr/~ahaury/
Analysis and Computational Dissection of Molecular Signature Multiplicity
Molecular signatures are computational or mathematical models created to diagnose disease and other phenotypes and to predict clinical outcomes and response to treatment. It is widely recognized that molecular signatures constitute one of the most important translational and basic science developments enabled by recent high-throughput molecular assays. A perplexing phenomenon that characterizes high-throughput data analysis is the ubiquitous multiplicity of molecular signatures. Multiplicity is a special form of data analysis instability in which different analysis methods used on the same data, or different samples from the same population lead to different but apparently maximally predictive signatures. This phenomenon has far-reaching implications for biological discovery and development of next generation patient diagnostics and personalized treatments. Currently the causes and interpretation of signature multiplicity are unknown, and several, often contradictory, conjectures have been made to explain it. We present a formal characterization of signature multiplicity and a new efficient algorithm that offers theoretical guarantees for extracting the set of maximally predictive and non-redundant signatures independent of distribution. The new algorithm identifies exactly the set of optimal signatures in controlled experiments and yields signatures with significantly better predictivity and reproducibility than previous algorithms in human microarray gene expression datasets. Our results shed light on the causes of signature multiplicity, provide computational tools for studying it empirically and introduce a framework for in silico bioequivalence of this important new class of diagnostic and personalized medicine modalities
Field theoretic approach to metastability in the contact process
A quantum field theoretic formulation of the dynamics of the Contact Process
on a regular graph of degree z is introduced. A perturbative calculation in
powers of 1/z of the effective potential for the density of particles phi(t)
and an instantonic field psi(t) emerging from the quantum formalism is
performed. Corrections to the mean-field distribution of densities of particles
in the out-of-equilibrium stationary state are derived in powers of 1/z.
Results for typical (e.g. average density) and rare fluctuation (e.g. lifetime
of the metastable state) properties are in very good agreement with numerical
simulations carried out on D-dimensional hypercubic (z=2D) and Cayley lattices.Comment: Final published version; 20 pages, 5 figure
A critical evaluation of network and pathway based classifiers for outcome prediction in breast cancer
Recently, several classifiers that combine primary tumor data, like gene
expression data, and secondary data sources, such as protein-protein
interaction networks, have been proposed for predicting outcome in breast
cancer. In these approaches, new composite features are typically constructed
by aggregating the expression levels of several genes. The secondary data
sources are employed to guide this aggregation. Although many studies claim
that these approaches improve classification performance over single gene
classifiers, the gain in performance is difficult to assess. This stems mainly
from the fact that different breast cancer data sets and validation procedures
are employed to assess the performance. Here we address these issues by
employing a large cohort of six breast cancer data sets as benchmark set and by
performing an unbiased evaluation of the classification accuracies of the
different approaches. Contrary to previous claims, we find that composite
feature classifiers do not outperform simple single gene classifiers. We
investigate the effect of (1) the number of selected features; (2) the specific
gene set from which features are selected; (3) the size of the training set and
(4) the heterogeneity of the data set on the performance of composite feature
and single gene classifiers. Strikingly, we find that randomization of
secondary data sources, which destroys all biological information in these
sources, does not result in a deterioration in performance of composite feature
classifiers. Finally, we show that when a proper correction for gene set size
is performed, the stability of single gene sets is similar to the stability of
composite feature sets. Based on these results there is currently no reason to
prefer prognostic classifiers based on composite features over single gene
classifiers for predicting outcome in breast cancer
- …