61 research outputs found
Toward Understanding Generative Data Augmentation
Generative data augmentation, which scales datasets by obtaining fake labeled
examples from a trained conditional generative model, boosts classification
performance in various learning tasks including (semi-)supervised learning,
few-shot learning, and adversarially robust learning. However, little work has
theoretically investigated the effect of generative data augmentation. To fill
this gap, we establish a general stability bound in this not independently and
identically distributed (non-i.i.d.) setting, where the learned distribution is
dependent on the original train set and generally not the same as the true
distribution. Our theoretical result includes the divergence between the
learned distribution and the true distribution. It shows that generative data
augmentation can enjoy a faster learning rate when the order of divergence term
is , where is the train
set size and is the corresponding stability constant. We further
specify the learning setup to the Gaussian mixture model and generative
adversarial nets. We prove that in both cases, though generative data
augmentation does not enjoy a faster learning rate, it can improve the learning
guarantees at a constant level when the train set is small, which is
significant when the awful overfitting occurs. Simulation results on the
Gaussian mixture model and empirical results on generative adversarial nets
support our theoretical conclusions. Our code is available at
https://github.com/ML-GSAI/Understanding-GDA.Comment: 39 page
EGSDE: Unpaired Image-to-Image Translation via Energy-Guided Stochastic Differential Equations
Score-based diffusion generative models (SDGMs) have achieved the SOTA FID
results in unpaired image-to-image translation (I2I). However, we notice that
existing methods totally ignore the training data in the source domain, leading
to sub-optimal solutions for unpaired I2I. To this end, we propose
energy-guided stochastic differential equations (EGSDE) that employs an energy
function pretrained on both the source and target domains to guide the
inference process of a pretrained SDE for realistic and faithful unpaired I2I.
Building upon two feature extractors, we carefully design the energy function
such that it encourages the transferred image to preserve the
domain-independent features and discard domainspecific ones. Further, we
provide an alternative explanation of the EGSDE as a product of experts, where
each of the three experts (corresponding to the SDE and two feature extractors)
solely contributes to faithfulness or realism. Empirically, we compare EGSDE to
a large family of baselines on three widely-adopted unpaired I2I tasks under
four metrics. EGSDE not only consistently outperforms existing SDGMs-based
methods in almost all settings but also achieves the SOTA realism results
(e.g., FID of 65.82 in Cat to Dog and FID of 59.75 in Wild to Dog on AFHQ)
without harming the faithful performance
- …