Manifold learning in Wasserstein space

Abstract

This paper aims at building the theoretical foundations for manifold learning algorithms in the space of absolutely continuous probability measures on a compact and convex subset of Rd\mathbb{R}^d, metrized with the Wasserstein-2 distance WW. We begin by introducing a natural construction of submanifolds Λ\Lambda of probability measures equipped with metric WΛW_\Lambda, the geodesic restriction of WW to Λ\Lambda. In contrast to other constructions, these submanifolds are not necessarily flat, but still allow for local linearizations in a similar fashion to Riemannian submanifolds of Rd\mathbb{R}^d. We then show how the latent manifold structure of (Λ,WΛ)(\Lambda,W_{\Lambda}) can be learned from samples {λi}i=1N\{\lambda_i\}_{i=1}^N of Λ\Lambda and pairwise extrinsic Wasserstein distances WW only. In particular, we show that the metric space (Λ,WΛ)(\Lambda,W_{\Lambda}) can be asymptotically recovered in the sense of Gromov--Wasserstein from a graph with nodes {λi}i=1N\{\lambda_i\}_{i=1}^N and edge weights W(λi,λj)W(\lambda_i,\lambda_j). In addition, we demonstrate how the tangent space at a sample λ\lambda can be asymptotically recovered via spectral analysis of a suitable "covariance operator" using optimal transport maps from λ\lambda to sufficiently close and diverse samples {λi}i=1N\{\lambda_i\}_{i=1}^N. The paper closes with some explicit constructions of submanifolds Λ\Lambda and numerical examples on the recovery of tangent spaces through spectral analysis

    Similar works

    Full text

    thumbnail-image

    Available Versions