5,667 research outputs found

    Learning Generative Models across Incomparable Spaces

    Full text link
    Generative Adversarial Networks have shown remarkable success in learning a distribution that faithfully recovers a reference distribution in its entirety. However, in some cases, we may want to only learn some aspects (e.g., cluster or manifold structure), while modifying others (e.g., style, orientation or dimension). In this work, we propose an approach to learn generative models across such incomparable spaces, and demonstrate how to steer the learned distribution towards target properties. A key component of our model is the Gromov-Wasserstein distance, a notion of discrepancy that compares distributions relationally rather than absolutely. While this framework subsumes current generative models in identically reproducing distributions, its inherent flexibility allows application to tasks in manifold learning, relational learning and cross-domain learning.Comment: International Conference on Machine Learning (ICML

    Coalgebraic Behavioral Metrics

    Get PDF
    We study different behavioral metrics, such as those arising from both branching and linear-time semantics, in a coalgebraic setting. Given a coalgebra α ⁣:XHX\alpha\colon X \to HX for a functor H ⁣:SetSetH \colon \mathrm{Set}\to \mathrm{Set}, we define a framework for deriving pseudometrics on XX which measure the behavioral distance of states. A crucial step is the lifting of the functor HH on Set\mathrm{Set} to a functor H\overline{H} on the category PMet\mathrm{PMet} of pseudometric spaces. We present two different approaches which can be viewed as generalizations of the Kantorovich and Wasserstein pseudometrics for probability measures. We show that the pseudometrics provided by the two approaches coincide on several natural examples, but in general they differ. If HH has a final coalgebra, every lifting H\overline{H} yields in a canonical way a behavioral distance which is usually branching-time, i.e., it generalizes bisimilarity. In order to model linear-time metrics (generalizing trace equivalences), we show sufficient conditions for lifting distributive laws and monads. These results enable us to employ the generalized powerset construction

    A new Poisson-type deviation inequality for Markov jump processes with positive Wasserstein curvature

    Full text link
    The purpose of this paper is to extend the investigation of Poisson-type deviation inequalities started by Joulin (Bernoulli 13 (2007) 782--798) to the empirical mean of positively curved Markov jump processes. In particular, our main result generalizes the tail estimates given by Lezaud (Ann. Appl. Probab. 8 (1998) 849--867, ESAIM Probab. Statist. 5 (2001) 183--201). An application to birth--death processes completes this work.Comment: Published in at http://dx.doi.org/10.3150/08-BEJ158 the Bernoulli (http://isi.cbs.nl/bernoulli/) by the International Statistical Institute/Bernoulli Society (http://isi.cbs.nl/BS/bshome.htm

    Manifold-valued Image Generation with Wasserstein Generative Adversarial Nets

    Full text link
    Generative modeling over natural images is one of the most fundamental machine learning problems. However, few modern generative models, including Wasserstein Generative Adversarial Nets (WGANs), are studied on manifold-valued images that are frequently encountered in real-world applications. To fill the gap, this paper first formulates the problem of generating manifold-valued images and exploits three typical instances: hue-saturation-value (HSV) color image generation, chromaticity-brightness (CB) color image generation, and diffusion-tensor (DT) image generation. For the proposed generative modeling problem, we then introduce a theorem of optimal transport to derive a new Wasserstein distance of data distributions on complete manifolds, enabling us to achieve a tractable objective under the WGAN framework. In addition, we recommend three benchmark datasets that are CIFAR-10 HSV/CB color images, ImageNet HSV/CB color images, UCL DT image datasets. On the three datasets, we experimentally demonstrate the proposed manifold-aware WGAN model can generate more plausible manifold-valued images than its competitors.Comment: Accepted by AAAI 201
    corecore