4,108 research outputs found
Large deviations in quantum lattice systems: one-phase region
We give large deviation upper bounds, and discuss lower bounds, for the
Gibbs-KMS state of a system of quantum spins or an interacting Fermi gas on the
lattice. We cover general interactions and general observables, both in the
high temperature regime and in dimension one.Comment: 30 pages, LaTeX 2
Ruelle-Lanford functions for quantum spin systems
We prove a large deviation principle for the expectation of macroscopic
observables in quantum (and classical) Gibbs states. Our proof is based on
Ruelle-Lanford functions and direct subadditivity arguments, as in the
classical case, instead of relying on G\"artner-Ellis theorem, and cluster
expansion or transfer operators as done in the quantum case. In this approach
we recover, expand, and unify quantum (and classical) large deviation results
for lattice Gibbs states. In the companion paper \cite{OR} we discuss the
characterization of rate functions in terms of relative entropies.Comment: 22 page
Decompositions of two player games: potential, zero-sum, and stable games
We introduce several methods of decomposition for two player normal form
games. Viewing the set of all games as a vector space, we exhibit explicit
orthonormal bases for the subspaces of potential games, zero-sum games, and
their orthogonal complements which we call anti-potential games and
anti-zero-sum games, respectively. Perhaps surprisingly, every anti-potential
game comes either from the Rock-Paper-Scissors type games (in the case of
symmetric games) or from the Matching Pennies type games (in the case of
asymmetric games). Using these decompositions, we prove old (and some new)
cycle criteria for potential and zero-sum games (as orthogonality relations
between subspaces). We illustrate the usefulness of our decomposition by (a)
analyzing the generalized Rock-Paper-Scissors game, (b) completely
characterizing the set of all null-stable games, (c) providing a large class of
strict stable games, (d) relating the game decomposition to the decomposition
of vector fields for the replicator equations, (e) constructing Lyapunov
functions for some replicator dynamics, and (f) constructing Zeeman games
-games with an interior asymptotically stable Nash equilibrium and a pure
strategy ESS
Scaling-up Empirical Risk Minimization: Optimization of Incomplete U-statistics
In a wide range of statistical learning problems such as ranking, clustering
or metric learning among others, the risk is accurately estimated by
-statistics of degree , i.e. functionals of the training data with
low variance that take the form of averages over -tuples. From a
computational perspective, the calculation of such statistics is highly
expensive even for a moderate sample size , as it requires averaging
terms. This makes learning procedures relying on the optimization of
such data functionals hardly feasible in practice. It is the major goal of this
paper to show that, strikingly, such empirical risks can be replaced by
drastically computationally simpler Monte-Carlo estimates based on terms
only, usually referred to as incomplete -statistics, without damaging the
learning rate of Empirical Risk Minimization (ERM)
procedures. For this purpose, we establish uniform deviation results describing
the error made when approximating a -process by its incomplete version under
appropriate complexity assumptions. Extensions to model selection, fast rate
situations and various sampling techniques are also considered, as well as an
application to stochastic gradient descent for ERM. Finally, numerical examples
are displayed in order to provide strong empirical evidence that the approach
we promote largely surpasses more naive subsampling techniques.Comment: To appear in Journal of Machine Learning Research. 34 pages. v2:
minor correction to Theorem 4 and its proof, added 1 reference. v3: typo
corrected in Proposition 3. v4: improved presentation, added experiments on
model selection for clustering, fixed minor typo
Similarity Learning for High-Dimensional Sparse Data
A good measure of similarity between data points is crucial to many tasks in
machine learning. Similarity and metric learning methods learn such measures
automatically from data, but they do not scale well respect to the
dimensionality of the data. In this paper, we propose a method that can learn
efficiently similarity measure from high-dimensional sparse data. The core idea
is to parameterize the similarity measure as a convex combination of rank-one
matrices with specific sparsity structures. The parameters are then optimized
with an approximate Frank-Wolfe procedure to maximally satisfy relative
similarity constraints on the training data. Our algorithm greedily
incorporates one pair of features at a time into the similarity measure,
providing an efficient way to control the number of active features and thus
reduce overfitting. It enjoys very appealing convergence guarantees and its
time and memory complexity depends on the sparsity of the data instead of the
dimension of the feature space. Our experiments on real-world high-dimensional
datasets demonstrate its potential for classification, dimensionality reduction
and data exploration.Comment: 14 pages. Proceedings of the 18th International Conference on
Artificial Intelligence and Statistics (AISTATS 2015). Matlab code:
https://github.com/bellet/HDS
- …