19,604 research outputs found

    Exploratory Analysis of Functional Data via Clustering and Optimal Segmentation

    Full text link
    We propose in this paper an exploratory analysis algorithm for functional data. The method partitions a set of functions into KK clusters and represents each cluster by a simple prototype (e.g., piecewise constant). The total number of segments in the prototypes, PP, is chosen by the user and optimally distributed among the clusters via two dynamic programming algorithms. The practical relevance of the method is shown on two real world datasets

    Improving the Convergence of Vector Fitting for Equivalent Circuit Extraction From Noisy Frequency Responses

    Get PDF
    The vector fitting (VF) algorithm has become a common tool in electromagnetic compatibility and signal integrity studies. This algorithm allows the derivation of a rational approximation to the transfer matrix of a given linear structure starting from measured or simulated frequency responses. This paper addresses the convergence properties of a VF when the frequency samples are affected by noise.We show that small amounts of noise can seriously impair or destroy convergence. This is due to the presence of spurious poles that appear during the iterations. To overcome this problem we suggest a simple modification of the basic VF algorithm, based on the identification and removal of the spurious poles. Also, an incremental pole addition and relocation process is proposed in order to provide automatic order estimation even in the presence of significant noise.We denote the resulting algorithm as vector fitting with adding and skimming (VF-AS). A thorough validation of the VF-AS algorithm is presented using a Monte Carlo analysis on synthetic noisy frequency responses. The results show excellent convergence and significant improvements with respect to the basic VF iteration scheme. Finally, we apply the new VF-AS algorithm to measured scattering responses of interconnect structures and networks typical of high-speed digital systems

    Supervised Learning in Multilayer Spiking Neural Networks

    Get PDF
    The current article introduces a supervised learning algorithm for multilayer spiking neural networks. The algorithm presented here overcomes some limitations of existing learning algorithms as it can be applied to neurons firing multiple spikes and it can in principle be applied to any linearisable neuron model. The algorithm is applied successfully to various benchmarks, such as the XOR problem and the Iris data set, as well as complex classifications problems. The simulations also show the flexibility of this supervised learning algorithm which permits different encodings of the spike timing patterns, including precise spike trains encoding.Comment: 38 pages, 4 figure

    Distributed Algorithms for Stochastic Source Seeking With Mobile Robot Networks

    Get PDF
    Autonomous robot networks are an effective tool for monitoring large-scale environmental fields. This paper proposes distributed control strategies for localizing the source of a noisy signal, which could represent a physical quantity of interest such as magnetic force, heat, radio signal, or chemical concentration. We develop algorithms specific to two scenarios: one in which the sensors have a precise model of the signal formation process and one in which a signal model is not available. In the model-free scenario, a team of sensors is used to follow a stochastic gradient of the signal field. Our approach is distributed, robust to deformations in the group geometry, does not necessitate global localization, and is guaranteed to lead the sensors to a neighborhood of a local maximum of the field. In the model-based scenario, the sensors follow a stochastic gradient of the mutual information (MI) between their expected measurements and the expected source location in a distributed manner. The performance is demonstrated in simulation using a robot sensor network to localize the source of a wireless radio signal

    Chebyshev Polynomial Approximation for Distributed Signal Processing

    Get PDF
    Unions of graph Fourier multipliers are an important class of linear operators for processing signals defined on graphs. We present a novel method to efficiently distribute the application of these operators to the high-dimensional signals collected by sensor networks. The proposed method features approximations of the graph Fourier multipliers by shifted Chebyshev polynomials, whose recurrence relations make them readily amenable to distributed computation. We demonstrate how the proposed method can be used in a distributed denoising task, and show that the communication requirements of the method scale gracefully with the size of the network.Comment: 8 pages, 5 figures, to appear in the Proceedings of the IEEE International Conference on Distributed Computing in Sensor Systems (DCOSS), June, 2011, Barcelona, Spai

    Dual Averaging for Distributed Optimization: Convergence Analysis and Network Scaling

    Full text link
    The goal of decentralized optimization over a network is to optimize a global objective formed by a sum of local (possibly nonsmooth) convex functions using only local computation and communication. It arises in various application domains, including distributed tracking and localization, multi-agent co-ordination, estimation in sensor networks, and large-scale optimization in machine learning. We develop and analyze distributed algorithms based on dual averaging of subgradients, and we provide sharp bounds on their convergence rates as a function of the network size and topology. Our method of analysis allows for a clear separation between the convergence of the optimization algorithm itself and the effects of communication constraints arising from the network structure. In particular, we show that the number of iterations required by our algorithm scales inversely in the spectral gap of the network. The sharpness of this prediction is confirmed both by theoretical lower bounds and simulations for various networks. Our approach includes both the cases of deterministic optimization and communication, as well as problems with stochastic optimization and/or communication.Comment: 40 pages, 4 figure
    corecore