1,901 research outputs found

    Applications of Recurrent Neural Networks to Optimization Problems

    Get PDF

    Synchronization and state estimation for discrete-time complex networks with distributed delays

    Get PDF
    Copyright [2008] IEEE. This material is posted here with permission of the IEEE. Such permission of the IEEE does not in any way imply IEEE endorsement of any of Brunel University's products or services. Internal or personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution must be obtained from the IEEE by writing to [email protected]. By choosing to view this document, you agree to all provisions of the copyright laws protecting it.In this paper, a synchronization problem is investigated for an array of coupled complex discrete-time networks with the simultaneous presence of both the discrete and distributed time delays. The complex networks addressed which include neural and social networks as special cases are quite general. Rather than the commonly used Lipschitz-type function, a more general sector-like nonlinear function is employed to describe the nonlinearities existing in the network. The distributed infinite time delays in the discrete-time domain are first defined. By utilizing a novel Lyapunov-Krasovskii functional and the Kronecker product, it is shown that the addressed discrete-time complex network with distributed delays is synchronized if certain linear matrix inequalities (LMIs) are feasible. The state estimation problem is then studied for the same complex network, where the purpose is to design a state estimator to estimate the network states through available output measurements such that, for all admissible discrete and distributed delays, the dynamics of the estimation error is guaranteed to be globally asymptotically stable. Again, an LMI approach is developed for the state estimation problem. Two simulation examples are provided to show the usefulness of the proposed global synchronization and state estimation conditions. It is worth pointing out that our main results are valid even if the nominal subsystems within the network are unstable

    Statistically optimum pre- and postfiltering in quantization

    Get PDF
    We consider the optimization of pre- and postfilters surrounding a quantization system. The goal is to optimize the filters such that the mean square error is minimized under the key constraint that the quantization noise variance is directly proportional to the variance of the quantization system input. Unlike some previous work, the postfilter is not restricted to be the inverse of the prefilter. With no order constraint on the filters, we present closed-form solutions for the optimum pre- and postfilters when the quantization system is a uniform quantizer. Using these optimum solutions, we obtain a coding gain expression for the system under study. The coding gain expression clearly indicates that, at high bit rates, there is no loss in generality in restricting the postfilter to be the inverse of the prefilter. We then repeat the same analysis with first-order pre- and postfilters in the form 1+αz-1 and 1/(1+γz^-1 ). In specific, we study two cases: 1) FIR prefilter, IIR postfilter and 2) IIR prefilter, FIR postfilter. For each case, we obtain a mean square error expression, optimize the coefficients α and γ and provide some examples where we compare the coding gain performance with the case of α=γ. In the last section, we assume that the quantization system is an orthonormal perfect reconstruction filter bank. To apply the optimum preand postfilters derived earlier, the output of the filter bank must be wide-sense stationary WSS which, in general, is not true. We provide two theorems, each under a different set of assumptions, that guarantee the wide sense stationarity of the filter bank output. We then propose a suboptimum procedure to increase the coding gain of the orthonormal filter bank

    A Brief Review on Mathematical Tools Applicable to Quantum Computing for Modelling and Optimization Problems in Engineering

    Get PDF
    Since its emergence, quantum computing has enabled a wide spectrum of new possibilities and advantages, including its efficiency in accelerating computational processes exponentially. This has directed much research towards completely novel ways of solving a wide variety of engineering problems, especially through describing quantum versions of many mathematical tools such as Fourier and Laplace transforms, differential equations, systems of linear equations, and optimization techniques, among others. Exploration and development in this direction will revolutionize the world of engineering. In this manuscript, we review the state of the art of these emerging techniques from the perspective of quantum computer development and performance optimization, with a focus on the most common mathematical tools that support engineering applications. This review focuses on the application of these mathematical tools to quantum computer development and performance improvement/optimization. It also identifies the challenges and limitations related to the exploitation of quantum computing and outlines the main opportunities for future contributions. This review aims at offering a valuable reference for researchers in fields of engineering that are likely to turn to quantum computing for solutions. Doi: 10.28991/ESJ-2023-07-01-020 Full Text: PD

    Time-stepping numerical simulation of switched circuits with the nonsmooth dynamical systems approach

    Get PDF
    International audienceThe numerical integration of switching circuits is known to be a tough issue when the number of switches is large, or when sliding modes exist. Then, classical analog simulators may behave poorly, or even fail. In this paper, it is shown on two examples that the nonsmooth dynamical systems (NSDS) approach, which is made of: 1) a specific modeling of the piecewise-linear electronic devices (ideal diodes, Zener diodes, transistors); 2) the Moreau's time-stepping scheme; and 3) specific iterative one-step solvers, supersedes simulators of the simulation program with integrated circuit emphasis (SPICE) family and hybrid simulators. An academic example constructed in [Maffezzoni, , IEEE Trans. CADICS, vol 25, no. 11, Nov. 2006], so that the Newton-Raphson scheme does not converge, and the buck converter are used to make extensive comparisons between the NSDS method and other methods of the SPICE family and a hybrid-like method. The NSDS method, implemented in the siconos platform developed at INRIA, proves to be on these two examples much faster and more robust with respect to the model parameter variations

    The nonsmooth dynamical systems approach for the analog simulation of switched circuits within the Siconos framework

    Get PDF
    The numerical integration of switching circuits is known to be a tough issue when the number of switches is high, or when sliding modes exist. Then classical analog simulators may behave poorly, or even fail. In this paper it is shown on two examples that the nonsmooth dynamical systems (NSDS) approach, which is made of 1) a specific modelling of the piecewise- linear electronic devices (ideal diodes, Zener diodes, transistors), 2) the Moreau's time-stepping scheme, and 3) specific iterative one-step solvers, supersedes simulators of the SPICE family and hybrid simulators. An academic example constructed in [Maffezzoni et al, IEEE Trans. on CADICS, Vol 25, No 11, November 2006], so that the Newton-Raphson scheme does not converge, and the buck converter, are used to make extensive comparisons between the NSDS method and other methods of the SPICE family and a hybrid-like method. The NSDS method, implemented in the Siconos platform developed at INRIA, proves to be on these two examples much faster and more robust with respect to the models parameters variations

    Supervised learning with quantum enhanced feature spaces

    Full text link
    Machine learning and quantum computing are two technologies each with the potential for altering how computation is performed to address previously untenable problems. Kernel methods for machine learning are ubiquitous for pattern recognition, with support vector machines (SVMs) being the most well-known method for classification problems. However, there are limitations to the successful solution to such problems when the feature space becomes large, and the kernel functions become computationally expensive to estimate. A core element to computational speed-ups afforded by quantum algorithms is the exploitation of an exponentially large quantum state space through controllable entanglement and interference. Here, we propose and experimentally implement two novel methods on a superconducting processor. Both methods represent the feature space of a classification problem by a quantum state, taking advantage of the large dimensionality of quantum Hilbert space to obtain an enhanced solution. One method, the quantum variational classifier builds on [1,2] and operates through using a variational quantum circuit to classify a training set in direct analogy to conventional SVMs. In the second, a quantum kernel estimator, we estimate the kernel function and optimize the classifier directly. The two methods present a new class of tools for exploring the applications of noisy intermediate scale quantum computers [3] to machine learning.Comment: Fixed typos, added figures and discussion about quantum error mitigatio
    • …
    corecore