1,109 research outputs found
Asymptotic behavior of memristive circuits
The interest in memristors has risen due to their possible application both
as memory units and as computational devices in combination with CMOS. This is
in part due to their nonlinear dynamics, and a strong dependence on the circuit
topology. We provide evidence that also purely memristive circuits can be
employed for computational purposes. In the present paper we show that a
polynomial Lyapunov function in the memory parameters exists for the case of DC
controlled memristors. Such Lyapunov function can be asymptotically
approximated with binary variables, and mapped to quadratic combinatorial
optimization problems. This also shows a direct parallel between memristive
circuits and the Hopfield-Little model. In the case of Erdos-Renyi random
circuits, we show numerically that the distribution of the matrix elements of
the projectors can be roughly approximated with a Gaussian distribution, and
that it scales with the inverse square root of the number of elements. This
provides an approximated but direct connection with the physics of disordered
system and, in particular, of mean field spin glasses. Using this and the fact
that the interaction is controlled by a projector operator on the loop space of
the circuit. We estimate the number of stationary points of the approximate
Lyapunov function and provide a scaling formula as an upper bound in terms of
the circuit topology only.Comment: 20 pages, 8 figures; proofs corrected, figures changed; results
substantially unchanged; to appear in Entrop
Reduction of dimension for nonlinear dynamical systems
We consider reduction of dimension for nonlinear dynamical systems. We
demonstrate that in some cases, one can reduce a nonlinear system of equations
into a single equation for one of the state variables, and this can be useful
for computing the solution when using a variety of analytical approaches. In
the case where this reduction is possible, we employ differential elimination
to obtain the reduced system. While analytical, the approach is algorithmic,
and is implemented in symbolic software such as {\sc MAPLE} or {\sc SageMath}.
In other cases, the reduction cannot be performed strictly in terms of
differential operators, and one obtains integro-differential operators, which
may still be useful. In either case, one can use the reduced equation to both
approximate solutions for the state variables and perform chaos diagnostics
more efficiently than could be done for the original higher-dimensional system,
as well as to construct Lyapunov functions which help in the large-time study
of the state variables. A number of chaotic and hyperchaotic dynamical systems
are used as examples in order to motivate the approach.Comment: 16 pages, no figure
Nonlinear system identification and control using dynamic multi-time scales neural networks
In this thesis, on-line identification algorithm and adaptive control design are proposed for nonlinear singularly perturbed systems which are represented by dynamic neural network model with multi-time scales. A novel on-line identification law for the Neural Network weights and linear part matrices of the model has been developed to minimize the identification errors. Based on the identification results, an adaptive controller is developed to achieve trajectory tracking. The Lyapunov synthesis method is used to conduct stability analysis for both identification algorithm and control design. To further enhance the stability and performance of the control system, an improved . dynamic neural network model is proposed by replacing all the output signals from the plant with the state variables of the neural network. Accordingly, the updating laws are modified with a dead-zone function to prevent parameter drifting. By combining feedback linearization with one of three classical control methods such as direct compensator, sliding mode controller or energy function compensation scheme, three different adaptive controllers have been proposed for trajectory tracking. New Lyapunov function analysis method is applied for the stability analysis of the improved identification algorithm and three control systems. Extensive simulation results are provided to support the effectiveness of the proposed identification algorithms and control systems for both dynamic NN models
New Stability Criterion for Takagi-Sugeno Fuzzy Cohen-Grossberg Neural Networks with Probabilistic Time-Varying Delays
A new global asymptotic stability criterion of Takagi-Sugeno fuzzy Cohen-Grossberg neural networks with probabilistic time-varying delays was derived, in which the diffusion item can play its role. Owing to deleting the boundedness conditions on amplification functions, the main result is a novelty to some extent. Besides, there is another novelty in methods, for Lyapunov-Krasovskii functional is the positive definite form of p powers, which is different from those of existing literature. Moreover, a numerical example illustrates the effectiveness of the proposed methods
Exploiting Nonlinear Recurrence and Fractal Scaling Properties for Voice Disorder Detection
Background: Voice disorders affect patients profoundly, and acoustic tools can potentially measure voice function objectively. Disordered sustained vowels exhibit wide-ranging phenomena, from nearly periodic to highly complex, aperiodic vibrations, and increased "breathiness". Modelling and surrogate data studies have shown significant nonlinear and non-Gaussian random properties in these sounds. Nonetheless, existing tools are limited to analysing voices displaying near periodicity, and do not account for this inherent biophysical nonlinearity and non-Gaussian randomness, often using linear signal processing methods insensitive to these properties. They do not directly measure the two main biophysical symptoms of disorder: complex nonlinear aperiodicity, and turbulent, aeroacoustic, non-Gaussian randomness. Often these tools cannot be applied to more severe disordered voices, limiting their clinical usefulness.

Methods: This paper introduces two new tools to speech analysis: recurrence and fractal scaling, which overcome the range limitations of existing tools by addressing directly these two symptoms of disorder, together reproducing a "hoarseness" diagram. A simple bootstrapped classifier then uses these two features to distinguish normal from disordered voices.

Results: On a large database of subjects with a wide variety of voice disorders, these new techniques can distinguish normal from disordered cases, using quadratic discriminant analysis, to overall correct classification performance of 91.8% plus or minus 2.0%. The true positive classification performance is 95.4% plus or minus 3.2%, and the true negative performance is 91.5% plus or minus 2.3% (95% confidence). This is shown to outperform all combinations of the most popular classical tools.

Conclusions: Given the very large number of arbitrary parameters and computational complexity of existing techniques, these new techniques are far simpler and yet achieve clinically useful classification performance using only a basic classification technique. They do so by exploiting the inherent nonlinearity and turbulent randomness in disordered voice signals. They are widely applicable to the whole range of disordered voice phenomena by design. These new measures could therefore be used for a variety of practical clinical purposes.

Neural-network dedicated processor for solving competitive assignment problems
A neural-network processor for solving first-order competitive assignment problems consists of a matrix of N x M processing units, each of which corresponds to the pairing of a first number of elements of (R sub i) with a second number of elements (C sub j), wherein limits of the first number are programmed in row control superneurons, and limits of the second number are programmed in column superneurons as MIN and MAX values. The cost (weight) W sub ij of the pairings is programmed separately into each PU. For each row and column of PU's, a dedicated constraint superneuron insures that the number of active neurons within the associated row or column fall within a specified range. Annealing is provided by gradually increasing the PU gain for each row and column or increasing positive feedback to each PU, the latter being effective to increase hysteresis of each PU or by combining both of these techniques
Minimizing Finite Sums with the Stochastic Average Gradient
We propose the stochastic average gradient (SAG) method for optimizing the
sum of a finite number of smooth convex functions. Like stochastic gradient
(SG) methods, the SAG method's iteration cost is independent of the number of
terms in the sum. However, by incorporating a memory of previous gradient
values the SAG method achieves a faster convergence rate than black-box SG
methods. The convergence rate is improved from O(1/k^{1/2}) to O(1/k) in
general, and when the sum is strongly-convex the convergence rate is improved
from the sub-linear O(1/k) to a linear convergence rate of the form O(p^k) for
p \textless{} 1. Further, in many cases the convergence rate of the new method
is also faster than black-box deterministic gradient methods, in terms of the
number of gradient evaluations. Numerical experiments indicate that the new
algorithm often dramatically outperforms existing SG and deterministic gradient
methods, and that the performance may be further improved through the use of
non-uniform sampling strategies.Comment: Revision from January 2015 submission. Major changes: updated
literature follow and discussion of subsequent work, additional Lemma showing
the validity of one of the formulas, somewhat simplified presentation of
Lyapunov bound, included code needed for checking proofs rather than the
polynomials generated by the code, added error regions to the numerical
experiment
- …