82,029 research outputs found
DNA Computing by Self-Assembly
Information and algorithms appear to be central to biological organization
and processes, from the storage and reproduction of genetic information to
the control of developmental processes to the sophisticated computations
performed by the nervous system. Much as human technology uses electronic
microprocessors to control electromechanical devices, biological
organisms use biochemical circuits to control molecular and chemical events.
The engineering and programming of biochemical circuits, in vivo and in
vitro, would transform industries that use chemical and nanostructured
materials. Although the construction of biochemical circuits has been
explored theoretically since the birth of molecular biology, our practical
experience with the capabilities and possible programming of biochemical
algorithms is still very young
"Going back to our roots": second generation biocomputing
Researchers in the field of biocomputing have, for many years, successfully
"harvested and exploited" the natural world for inspiration in developing
systems that are robust, adaptable and capable of generating novel and even
"creative" solutions to human-defined problems. However, in this position paper
we argue that the time has now come for a reassessment of how we exploit
biology to generate new computational systems. Previous solutions (the "first
generation" of biocomputing techniques), whilst reasonably effective, are crude
analogues of actual biological systems. We believe that a new, inherently
inter-disciplinary approach is needed for the development of the emerging
"second generation" of bio-inspired methods. This new modus operandi will
require much closer interaction between the engineering and life sciences
communities, as well as a bidirectional flow of concepts, applications and
expertise. We support our argument by examining, in this new light, three
existing areas of biocomputing (genetic programming, artificial immune systems
and evolvable hardware), as well as an emerging area (natural genetic
engineering) which may provide useful pointers as to the way forward.Comment: Submitted to the International Journal of Unconventional Computin
Fast, accurate, and transferable many-body interatomic potentials by symbolic regression
The length and time scales of atomistic simulations are limited by the
computational cost of the methods used to predict material properties. In
recent years there has been great progress in the use of machine learning
algorithms to develop fast and accurate interatomic potential models, but it
remains a challenge to develop models that generalize well and are fast enough
to be used at extreme time and length scales. To address this challenge, we
have developed a machine learning algorithm based on symbolic regression in the
form of genetic programming that is capable of discovering accurate,
computationally efficient manybody potential models. The key to our approach is
to explore a hypothesis space of models based on fundamental physical
principles and select models within this hypothesis space based on their
accuracy, speed, and simplicity. The focus on simplicity reduces the risk of
overfitting the training data and increases the chances of discovering a model
that generalizes well. Our algorithm was validated by rediscovering an exact
Lennard-Jones potential and a Sutton Chen embedded atom method potential from
training data generated using these models. By using training data generated
from density functional theory calculations, we found potential models for
elemental copper that are simple, as fast as embedded atom models, and capable
of accurately predicting properties outside of their training set. Our approach
requires relatively small sets of training data, making it possible to generate
training data using highly accurate methods at a reasonable computational cost.
We present our approach, the forms of the discovered models, and assessments of
their transferability, accuracy and speed
Searching for Globally Optimal Functional Forms for Inter-Atomic Potentials Using Parallel Tempering and Genetic Programming
We develop a Genetic Programming-based methodology that enables discovery of
novel functional forms for classical inter-atomic force-fields, used in
molecular dynamics simulations. Unlike previous efforts in the field, that fit
only the parameters to the fixed functional forms, we instead use a novel
algorithm to search the space of many possible functional forms. While a
follow-on practical procedure will use experimental and {\it ab inito} data to
find an optimal functional form for a forcefield, we first validate the
approach using a manufactured solution. This validation has the advantage of a
well-defined metric of success. We manufactured a training set of atomic
coordinate data with an associated set of global energies using the well-known
Lennard-Jones inter-atomic potential. We performed an automatic functional form
fitting procedure starting with a population of random functions, using a
genetic programming functional formulation, and a parallel tempering
Metropolis-based optimization algorithm. Our massively-parallel method
independently discovered the Lennard-Jones function after searching for several
hours on 100 processors and covering a miniscule portion of the configuration
space. We find that the method is suitable for unsupervised discovery of
functional forms for inter-atomic potentials/force-fields. We also find that
our parallel tempering Metropolis-based approach significantly improves the
optimization convergence time, and takes good advantage of the parallel cluster
architecture
Using rapid chlorophyll fluorescence transients to classify vitis genotypes
Silva, J. M. da, Figueiredo, A., Cunha, J., Eiras-Dias, J. E., Silva, S., Vanneschi, L., & Mariano, P. (2020). Using rapid chlorophyll fluorescence transients to classify vitis genotypes. Plants, 9(2), 1-19. [174]. https://doi.org/10.3390/plants9020174When a dark-adapted leaf is illuminated with saturating light, a fast polyphasic rise of fluorescence emission (Kautsky effect) is observed. The shape of the curve is dependent on the molecular organization of the photochemical apparatus, which in turn is a function of the interaction between genotype and environment. In this paper, we evaluate the potential of rapid fluorescence transients, aided by machine learning techniques, to classify plant genotypes. We present results of the application of several machine learning algorithms (k-nearest neighbors, decision trees, artificial neural networks, genetic programming) to rapid induction curves recorded in different species and cultivars of vine grown in the same environmental conditions. The phylogenetic relations between the selected Vitis species and Vitis vinifera cultivars were established with molecular markers. Both neural networks (71.8%) and genetic programming (75.3%) presented much higher global classification success rates than k-nearest neighbors (58.5%) or decision trees (51.6%), genetic programming performing slightly better than neural networks. However, compared with a random classifier (success rate = 14%), even the less successful algorithms were good at the task of classifying. The use of rapid fluorescence transients, handled by genetic programming, for rapid preliminary classification of Vitis genotypes is foreseen as feasible.publishersversionpublishe
Genetic programming in data mining for drug discovery
Genetic programming (GP) is used to extract from rat oral bioavailability
(OB) measurements simple, interpretable and predictive QSAR models
which both generalise to rats and to marketed drugs in humans. Receiver
Operating Characteristics (ROC) curves for the binary classier produced
by machine learning show no statistical dierence between rats (albeit
without known clearance dierences) and man. Thus evolutionary computing
oers the prospect of in silico ADME screening, e.g. for \virtual"
chemicals, for pharmaceutical drug discovery
Epigenetic inheritance. Concepts, mechanisms and perspectives
Parents' stressful experiences can influence an offspring's vulnerability to many pathological conditions, including psychopathologies, and their effects may even endure for several generations. Nevertheless, the cause of this phenomenon has not been determined, and only recently have scientists turned to epigenetics to answer this question. There is extensive literature on epigenetics, but no consensus exists with regard to how and what can (and must) be considered to study and define epigenetics processes and their inheritance. In this work, we aimed to clarify and systematize these concepts. To this end, we analyzed the dynamics of epigenetic changes over time in detail and defined three types of epigenetics: a direct form of epigenetics (DE) and two indirect epigenetic processes-within (WIE) and across (AIE). DE refers to changes that occur in the lifespan of an individual, due to direct experiences with his environment. WIE concerns changes that occur inside of the womb, due to events during gestation. Finally, AIE defines changes that affect the individual's predecessors (parents, grandparents, etc.), due to events that occur even long before conception and that are somehow (e.g., through gametes, the intrauterine environment setting) transmitted across generations. This distinction allows us to organize the main body of epigenetic evidence according to these categories and then focus on the latter (AIE), referring to it as a faster route of informational transmission across generations-compared with genetic inheritance-that guides human evolution in a Lamarckian (i.e., experience-dependent) manner. Of the molecular processes that are implicated in this phenomenon, well-known (methylation) and novel (non-coding RNA, ncRNA) regulatory mechanisms are converging. Our discussion of the chief methods that are used to study epigenetic inheritance highlights the most compelling technical and theoretical problems of this discipline. Experimental suggestions to expand this field are provided, and their practical and ethical implications are discussed extensivel
Neutrality: A Necessity for Self-Adaptation
Self-adaptation is used in all main paradigms of evolutionary computation to
increase efficiency. We claim that the basis of self-adaptation is the use of
neutrality. In the absence of external control neutrality allows a variation of
the search distribution without the risk of fitness loss.Comment: 6 pages, 3 figures, LaTe
- …