31 research outputs found
A novel quaternion linear matrix equation solver through zeroing neural networks with applications to acoustic source tracking
Due to its significance in science and engineering, time-varying linear matrix equation (LME) problems have received a lot of attention from scholars. It is for this reason that the issue of finding the minimum-norm least-squares solution of the time-varying quaternion LME (ML-TQ-LME) is addressed in this study. This is accomplished using the zeroing neural network (ZNN) technique, which has achieved considerable success in tackling time-varying issues. In light of that, two new ZNN models are introduced to solve the ML-TQ-LME problem for time-varying quaternion matrices of arbitrary dimension. Two simulation experiments and two practical acoustic source tracking applications show that the models function superbly
Complex Noise-Resistant Zeroing Neural Network for Computing Complex Time-Dependent Lyapunov Equation
Complex time-dependent Lyapunov equation (CTDLE), as an important means of stability analysis of control systems, has been extensively employed in mathematics and engineering application fields. Recursive neural networks (RNNs) have been reported as an effective method for solving CTDLE. In the previous work, zeroing neural networks (ZNNs) have been established to find the accurate solution of time-dependent Lyapunov equation (TDLE) in the noise-free conditions. However, noises are inevitable in the actual implementation process. In order to suppress the interference of various noises in practical applications, in this paper, a complex noise-resistant ZNN (CNRZNN) model is proposed and employed for the CTDLE solution. Additionally, the convergence and robustness of the CNRZNN model are analyzed and proved theoretically. For verification and comparison, three experiments and the existing noise-tolerant ZNN (NTZNN) model are introduced to investigate the effectiveness, convergence and robustness of the CNRZNN model. Compared with the NTZNN model, the CNRZNN model has more generality and stronger robustness. Specifically, the NTZNN model is a special form of the CNRZNN model, and the residual error of CNRZNN can converge rapidly and stably to order 10−5 when solving CTDLE under complex linear noises, which is much lower than order 10−1 of the NTZNN model. Analogously, under complex quadratic noises, the residual error of the CNRZNN model can converge to 2∥A∥F/ζ3 quickly and stably, while the residual error of the NTZNN model is divergent
Solving quaternion nonsymmetric algebraic Riccati equations through zeroing neural networks
Many variations of the algebraic Riccati equation (ARE) have been used to study nonlinear system stability in the control domain in great detail. Taking the quaternion nonsymmetric ARE (QNARE) as a generalized version of ARE, the time-varying QNARE (TQNARE) is introduced. This brings us to the main objective of this work: finding the TQNARE solution. The zeroing neural network (ZNN) technique, which has demonstrated a high degree of effectiveness in handling time-varying problems, is used to do this. Specifically, the TQNARE can be solved using the high order ZNN (HZNN) design, which is a member of the family of ZNN models that correlate to hyperpower iterative techniques. As a result, a novel HZNN model, called HZ-QNARE, is presented for solving the TQNARE. The model functions fairly well, as demonstrated by two simulation tests. Additionally, the results demonstrated that, while both approaches function remarkably well, the HZNN architecture works better than the ZNN architecture
Recurrent neural networks for solving matrix algebra problems
The aim of this dissertation is the application of recurrent neural
networks (RNNs) to solving some problems from a matrix algebra
with particular reference to the computations of the generalized
inverses as well as solving the matrix equations of constant (timeinvariant)
matrices. We examine the ability to exploit the correlation
between the dynamic state equations of recurrent neural networks for
computing generalized inverses and integral representations of these
generalized inverses. Recurrent neural networks are composed of
independent parts (sub-networks). These sub-networks can work
simultaneously, so parallel and distributed processing can be
accomplished. In this way, the computational advantages over the
existing sequential algorithms can be attained in real-time
applications. We investigate and exploit an analogy between the
scaled hyperpower family (SHPI family) of iterative methods for
computing the matrix inverse and the discretization of Zhang Neural
Network (ZNN) models. A class of ZNN models corresponding to the
family of hyperpower iterative methods for computing the generalized
inverses on the basis of the discovered analogy is defined. The Matlab
Simulink implementation of the introduced ZNN models is described
in the case of scaled hyperpower methods of the order 2 and 3. We
present the Matlab Simulink model of a hybrid recursive neural
implicit dynamics and give a simulation and comparison to the
existing Zhang dynamics for real-time matrix inversion. Simulation
results confirm a superior convergence of the hybrid model compared
to Zhang model
Recommended from our members
Efficient Predefined-Time Adaptive Neural Networks for Computing Time-Varying Tensor Moore–Penrose Inverse
This article proposes predefined-time adaptive neural network (PTANN) and event-triggered PTANN (ET-PTANN) models to efficiently compute the time-varying tensor Moore–Penrose (MP) inverse. The PTANN model incorporates a novel adaptive parameter and activation function, enabling it to achieve strongly predefined-time convergence. Unlike traditional time-varying parameters that increase over time, the adaptive parameter is proportional to the error norm, thereby better allocating computational resources and improving efficiency. To further enhance efficiency, the ET-PTANN model combines an event trigger with the evolution formula, resulting in the adjustment of step size and reduction of computation frequency compared to the PTANN model. By conducting mathematical derivations, the article derives the upper bound of convergence time for the proposed neural network models and determines the minimum execution interval for the event trigger. A simulation example demonstrates that the PTANN and ET-PTANN models outperform other related neural network models in terms of computational efficiency and convergence rate. Finally, the practicality of the PTANN and ET-PTANN models is demonstrated through their application for mobile sound source localization.Natural Science Foundation of Hunan Province of China (Grant Number: 2022RC1103 and 2021JJ20005);
10.13039/501100001809-National Natural Science Foundation of China (Grant Number: 61866013
Recommended from our members
Inference of functional neural connectivity and convergence acceleration methods
The knowledge of the maps of neuronal interactions is key for system neuroscience, but at the moment we possess relatively little of it . The recent development of experimental methods which allow a simultaneous recording of the spiking activity, but not the intracellular voltage, of thousands of neurons gives us an opportunity to start filling that gap. In Chapter 2, I present a method for the inference of the parameters of the leaky integrate-and-fire (LIF) model featuring time-dependent currents and conductances based only on the extracellular recording of spiking in the network. The fitted parameters can describe the functional connections in the network, as well as the internal properties of the cells. The method can also be used to determine whether a single-compartment model of a neuron should include conductance- or current-based synapses, or their mixture. In addition, because the same mathematical model describes some of the flavors of the Drift Diffusion Model (DDM), popular in the studies of decision making process, the presented method can be readily used to fit their parameters. Making the proposed inference procedure -- based on the expectation-maximization (EM) algorithm -- accurate and robust, necessitated a development of a new numerical adaptive-grid (AG) method for the forward-backward (FB) propagation of the probability density, which is required in the computation of the sufficient statistic in the EM algorithm. These topics are covered in Chapter 3. Another issue which had to be addressed in order to obtain a usable inference algorithm is the well known slow convergence of the EM algorithm in the flat regions of the loglikelihood. Two complementary approaches to this issue are presented in this dissertation. In Chapter 4, I present a new framework for the acceleration of convergence of iterative algorithms (not limited to the EM) which unifies all previously known methods and allows us to construct a new method demonstrating the best performance of them all. To make the computations even faster, I wrote a Matlab package which allows them to be done in parallel on several machines and clusters. As one can see, all the aforementioned projects were sprouted up from one "head" project on the inference of the LIF model parameters. At the end of the dissertation, I briefly describe a disconnected project which is devoted to the development of a flexible experimental setup (software and hardware) for behavioral experiments, with a specific application to a particular type of the virtual Morris water maze experiment (VMWM)
Statistical Analysis of Networks
This book is a general introduction to the statistical analysis of networks, and can serve both as a research monograph and as a textbook. Numerous fundamental tools and concepts needed for the analysis of networks are presented, such as network modeling, community detection, graph-based semi-supervised learning and sampling in networks. The description of these concepts is self-contained, with both theoretical justifications and applications provided for the presented algorithms.
Researchers, including postgraduate students, working in the area of network science, complex network analysis, or social network analysis, will find up-to-date statistical methods relevant to their research tasks. This book can also serve as textbook material for courses related to the
statistical approach to the analysis of complex networks.
In general, the chapters are fairly independent and self-supporting, and the book could be used for course composition “à la carte”. Nevertheless, Chapter 2 is needed to a certain degree for all parts of the book. It is also recommended to read Chapter 4 before reading Chapters 5 and 6, but this is not absolutely necessary. Reading Chapter 3 can also be helpful before reading Chapters 5 and 7.
As prerequisites for reading this book, a basic knowledge in probability, linear algebra and elementary notions of graph theory is advised. Appendices describing required notions from the above mentioned disciplines have been added to help readers gain further understanding