552 research outputs found
Scaling Algorithms for Unbalanced Transport Problems
This article introduces a new class of fast algorithms to approximate
variational problems involving unbalanced optimal transport. While classical
optimal transport considers only normalized probability distributions, it is
important for many applications to be able to compute some sort of relaxed
transportation between arbitrary positive measures. A generic class of such
"unbalanced" optimal transport problems has been recently proposed by several
authors. In this paper, we show how to extend the, now classical, entropic
regularization scheme to these unbalanced problems. This gives rise to fast,
highly parallelizable algorithms that operate by performing only diagonal
scaling (i.e. pointwise multiplications) of the transportation couplings. They
are generalizations of the celebrated Sinkhorn algorithm. We show how these
methods can be used to solve unbalanced transport, unbalanced gradient flows,
and to compute unbalanced barycenters. We showcase applications to 2-D shape
modification, color transfer, and growth models
Regularized Optimal Transport and the Rot Mover's Distance
This paper presents a unified framework for smooth convex regularization of
discrete optimal transport problems. In this context, the regularized optimal
transport turns out to be equivalent to a matrix nearness problem with respect
to Bregman divergences. Our framework thus naturally generalizes a previously
proposed regularization based on the Boltzmann-Shannon entropy related to the
Kullback-Leibler divergence, and solved with the Sinkhorn-Knopp algorithm. We
call the regularized optimal transport distance the rot mover's distance in
reference to the classical earth mover's distance. We develop two generic
schemes that we respectively call the alternate scaling algorithm and the
non-negative alternate scaling algorithm, to compute efficiently the
regularized optimal plans depending on whether the domain of the regularizer
lies within the non-negative orthant or not. These schemes are based on
Dykstra's algorithm with alternate Bregman projections, and further exploit the
Newton-Raphson method when applied to separable divergences. We enhance the
separable case with a sparse extension to deal with high data dimensions. We
also instantiate our proposed framework and discuss the inherent specificities
for well-known regularizers and statistical divergences in the machine learning
and information geometry communities. Finally, we demonstrate the merits of our
methods with experiments using synthetic data to illustrate the effect of
different regularizers and penalties on the solutions, as well as real-world
data for a pattern recognition application to audio scene classification
Inverse optimal transport
Discrete optimal transportation problems arise in various contexts in
engineering, the sciences and the social sciences. Often the underlying cost
criterion is unknown, or only partly known, and the observed optimal solutions
are corrupted by noise. In this paper we propose a systematic approach to infer
unknown costs from noisy observations of optimal transportation plans. The
algorithm requires only the ability to solve the forward optimal transport
problem, which is a linear program, and to generate random numbers. It has a
Bayesian interpretation, and may also be viewed as a form of stochastic
optimization.
We illustrate the developed methodologies using the example of international
migration flows. Reported migration flow data captures (noisily) the number of
individuals moving from one country to another in a given period of time. It
can be interpreted as a noisy observation of an optimal transportation map,
with costs related to the geographical position of countries. We use a
graph-based formulation of the problem, with countries at the nodes of graphs
and non-zero weighted adjacencies only on edges between countries which share a
border. We use the proposed algorithm to estimate the weights, which represent
cost of transition, and to quantify uncertainty in these weights
Quadratically-Regularized Optimal Transport on Graphs
Optimal transportation provides a means of lifting distances between points
on a geometric domain to distances between signals over the domain, expressed
as probability distributions. On a graph, transportation problems can be used
to express challenging tasks involving matching supply to demand with minimal
shipment expense; in discrete language, these become minimum-cost network flow
problems. Regularization typically is needed to ensure uniqueness for the
linear ground distance case and to improve optimization convergence;
state-of-the-art techniques employ entropic regularization on the
transportation matrix. In this paper, we explore a quadratic alternative to
entropic regularization for transport over a graph. We theoretically analyze
the behavior of quadratically-regularized graph transport, characterizing how
regularization affects the structure of flows in the regime of small but
nonzero regularization. We further exploit elegant second-order structure in
the dual of this problem to derive an easily-implemented Newton-type
optimization algorithm.Comment: 27 page
Learning Generative Models across Incomparable Spaces
Generative Adversarial Networks have shown remarkable success in learning a
distribution that faithfully recovers a reference distribution in its entirety.
However, in some cases, we may want to only learn some aspects (e.g., cluster
or manifold structure), while modifying others (e.g., style, orientation or
dimension). In this work, we propose an approach to learn generative models
across such incomparable spaces, and demonstrate how to steer the learned
distribution towards target properties. A key component of our model is the
Gromov-Wasserstein distance, a notion of discrepancy that compares
distributions relationally rather than absolutely. While this framework
subsumes current generative models in identically reproducing distributions,
its inherent flexibility allows application to tasks in manifold learning,
relational learning and cross-domain learning.Comment: International Conference on Machine Learning (ICML
- …