71,897 research outputs found
Quantum Genetics, Quantum Automata and Quantum Computation
The concepts of quantum automata and quantum computation are studied in the context of quantum genetics and genetic networks with nonlinear dynamics. In a previous publication (Baianu,1971a) the formal concept of quantum automaton was introduced and its possible implications for genetic and metabolic activities in living cells and organisms were considered. This was followed by a report on quantum and abstract, symbolic computation based on the theory of categories, functors and natural transformations (Baianu,1971b). The notions of topological semigroup, quantum automaton,or quantum computer, were then suggested with a view to their potential applications to the analogous simulation of biological systems, and especially genetic activities and nonlinear dynamics in genetic networks. Further, detailed studies of nonlinear dynamics in genetic networks were carried out in categories of n-valued, Lukasiewicz Logic Algebras that showed significant dissimilarities (Baianu, 1977) from Bolean models of human neural networks (McCullough and Pitts,1945). Molecular models in terms of categories, functors and natural transformations were then formulated for uni-molecular chemical transformations, multi-molecular chemical and biochemical transformations (Baianu, 1983,2004a). Previous applications of computer modeling, classical automata theory, and relational biology to molecular biology, oncogenesis and medicine were extensively reviewed and several important conclusions were reached regarding both the potential and limitations of the computation-assisted modeling of biological systems, and especially complex organisms such as Homo sapiens sapiens(Baianu,1987). Novel approaches to solving the realization problems of Relational Biology models in Complex System Biology are introduced in terms of natural transformations between functors of such molecular categories. Several applications of such natural transformations of functors were then presented to protein biosynthesis, embryogenesis and nuclear transplant experiments. Other possible realizations in Molecular Biology and Relational Biology of Organisms are here suggested in terms of quantum automata models of Quantum Genetics and Interactomics. Future developments of this novel approach are likely to also include: Fuzzy Relations in Biology and Epigenomics, Relational Biology modeling of Complex Immunological and Hormonal regulatory systems, n-categories and Topoi of Lukasiewicz Logic Algebras and Intuitionistic Logic (Heyting) Algebras for modeling nonlinear dynamics and cognitive processes in complex neural networks that are present in the human brain, as well as stochastic modeling of genetic networks in Lukasiewicz Logic Algebras
Universal neural field computation
Turing machines and G\"odel numbers are important pillars of the theory of
computation. Thus, any computational architecture needs to show how it could
relate to Turing machines and how stable implementations of Turing computation
are possible. In this chapter, we implement universal Turing computation in a
neural field environment. To this end, we employ the canonical symbologram
representation of a Turing machine obtained from a G\"odel encoding of its
symbolic repertoire and generalized shifts. The resulting nonlinear dynamical
automaton (NDA) is a piecewise affine-linear map acting on the unit square that
is partitioned into rectangular domains. Instead of looking at point dynamics
in phase space, we then consider functional dynamics of probability
distributions functions (p.d.f.s) over phase space. This is generally described
by a Frobenius-Perron integral transformation that can be regarded as a neural
field equation over the unit square as feature space of a dynamic field theory
(DFT). Solving the Frobenius-Perron equation yields that uniform p.d.f.s with
rectangular support are mapped onto uniform p.d.f.s with rectangular support,
again. We call the resulting representation \emph{dynamic field automaton}.Comment: 21 pages; 6 figures. arXiv admin note: text overlap with
arXiv:1204.546
Sampling-Based Nonlinear MPC of Neural Network Dynamics with Application to Autonomous Vehicle Motion Planning
Control of machine learning models has emerged as an important paradigm for a
broad range of robotics applications. In this paper, we present a
sampling-based nonlinear model predictive control (NMPC) approach for control
of neural network dynamics. We show its design in two parts: 1) formulating
conventional optimization-based NMPC as a Bayesian state estimation problem,
and 2) using particle filtering/smoothing to achieve the estimation. Through a
principled sampling-based implementation, this approach can potentially make
effective searches in the control action space for optimal control and also
facilitate computation toward overcoming the challenges caused by neural
network dynamics. We apply the proposed NMPC approach to motion planning for
autonomous vehicles. The specific problem considers nonlinear unknown vehicle
dynamics modeled as neural networks as well as dynamic on-road driving
scenarios. The approach shows significant effectiveness in successful motion
planning in case studies.Comment: To appear in 2022 American Control Conference (ACC
A Comparative Study of Reservoir Computing for Temporal Signal Processing
Reservoir computing (RC) is a novel approach to time series prediction using
recurrent neural networks. In RC, an input signal perturbs the intrinsic
dynamics of a medium called a reservoir. A readout layer is then trained to
reconstruct a target output from the reservoir's state. The multitude of RC
architectures and evaluation metrics poses a challenge to both practitioners
and theorists who study the task-solving performance and computational power of
RC. In addition, in contrast to traditional computation models, the reservoir
is a dynamical system in which computation and memory are inseparable, and
therefore hard to analyze. Here, we compare echo state networks (ESN), a
popular RC architecture, with tapped-delay lines (DL) and nonlinear
autoregressive exogenous (NARX) networks, which we use to model systems with
limited computation and limited memory respectively. We compare the performance
of the three systems while computing three common benchmark time series:
H{\'e}non Map, NARMA10, and NARMA20. We find that the role of the reservoir in
the reservoir computing paradigm goes beyond providing a memory of the past
inputs. The DL and the NARX network have higher memorization capability, but
fall short of the generalization power of the ESN
Product Reservoir Computing: Time-Series Computation with Multiplicative Neurons
Echo state networks (ESN), a type of reservoir computing (RC) architecture,
are efficient and accurate artificial neural systems for time series processing
and learning. An ESN consists of a core of recurrent neural networks, called a
reservoir, with a small number of tunable parameters to generate a
high-dimensional representation of an input, and a readout layer which is
easily trained using regression to produce a desired output from the reservoir
states. Certain computational tasks involve real-time calculation of high-order
time correlations, which requires nonlinear transformation either in the
reservoir or the readout layer. Traditional ESN employs a reservoir with
sigmoid or tanh function neurons. In contrast, some types of biological neurons
obey response curves that can be described as a product unit rather than a sum
and threshold. Inspired by this class of neurons, we introduce a RC
architecture with a reservoir of product nodes for time series computation. We
find that the product RC shows many properties of standard ESN such as
short-term memory and nonlinear capacity. On standard benchmarks for chaotic
prediction tasks, the product RC maintains the performance of a standard
nonlinear ESN while being more amenable to mathematical analysis. Our study
provides evidence that such networks are powerful in highly nonlinear tasks
owing to high-order statistics generated by the recurrent product node
reservoir
Memristors for the Curious Outsiders
We present both an overview and a perspective of recent experimental advances
and proposed new approaches to performing computation using memristors. A
memristor is a 2-terminal passive component with a dynamic resistance depending
on an internal parameter. We provide an brief historical introduction, as well
as an overview over the physical mechanism that lead to memristive behavior.
This review is meant to guide nonpractitioners in the field of memristive
circuits and their connection to machine learning and neural computation.Comment: Perpective paper for MDPI Technologies; 43 page
- …