8,238 research outputs found
A Framework for Algorithm Stability
We say that an algorithm is stable if small changes in the input result in
small changes in the output. This kind of algorithm stability is particularly
relevant when analyzing and visualizing time-varying data. Stability in general
plays an important role in a wide variety of areas, such as numerical analysis,
machine learning, and topology, but is poorly understood in the context of
(combinatorial) algorithms. In this paper we present a framework for analyzing
the stability of algorithms. We focus in particular on the tradeoff between the
stability of an algorithm and the quality of the solution it computes. Our
framework allows for three types of stability analysis with increasing degrees
of complexity: event stability, topological stability, and Lipschitz stability.
We demonstrate the use of our stability framework by applying it to kinetic
Euclidean minimum spanning trees
Computational Topology Techniques for Characterizing Time-Series Data
Topological data analysis (TDA), while abstract, allows a characterization of
time-series data obtained from nonlinear and complex dynamical systems. Though
it is surprising that such an abstract measure of structure - counting pieces
and holes - could be useful for real-world data, TDA lets us compare different
systems, and even do membership testing or change-point detection. However, TDA
is computationally expensive and involves a number of free parameters. This
complexity can be obviated by coarse-graining, using a construct called the
witness complex. The parametric dependence gives rise to the concept of
persistent homology: how shape changes with scale. Its results allow us to
distinguish time-series data from different systems - e.g., the same note
played on different musical instruments.Comment: 12 pages, 6 Figures, 1 Table, The Sixteenth International Symposium
on Intelligent Data Analysis (IDA 2017
Invariances and Data Augmentation for Supervised Music Transcription
This paper explores a variety of models for frame-based music transcription,
with an emphasis on the methods needed to reach state-of-the-art on human
recordings. The translation-invariant network discussed in this paper, which
combines a traditional filterbank with a convolutional neural network, was the
top-performing model in the 2017 MIREX Multiple Fundamental Frequency
Estimation evaluation. This class of models shares parameters in the
log-frequency domain, which exploits the frequency invariance of music to
reduce the number of model parameters and avoid overfitting to the training
data. All models in this paper were trained with supervision by labeled data
from the MusicNet dataset, augmented by random label-preserving pitch-shift
transformations.Comment: 6 page
Evolutionary Approaches to Minimizing Network Coding Resources
We wish to minimize the resources used for network coding while achieving the
desired throughput in a multicast scenario. We employ evolutionary approaches,
based on a genetic algorithm, that avoid the computational complexity that
makes the problem NP-hard. Our experiments show great improvements over the
sub-optimal solutions of prior methods. Our new algorithms improve over our
previously proposed algorithm in three ways. First, whereas the previous
algorithm can be applied only to acyclic networks, our new method works also
with networks with cycles. Second, we enrich the set of components used in the
genetic algorithm, which improves the performance. Third, we develop a novel
distributed framework. Combining distributed random network coding with our
distributed optimization yields a network coding protocol where the resources
used for coding are optimized in the setup phase by running our evolutionary
algorithm at each node of the network. We demonstrate the effectiveness of our
approach by carrying out simulations on a number of different sets of network
topologies.Comment: 9 pages, 6 figures, accepted to the 26th Annual IEEE Conference on
Computer Communications (INFOCOM 2007
- …