1,772 research outputs found
Perceptual Adversarial Networks With a Feature Pyramid for Image Translation
This paper investigates the image-to-image translations problems, where the input image is translated into its synthetic form with the original structure and semantics preserved. Widely used methods compute the pixel-wise MSE loss, which are often inadequate for high-frequency content and tend to produce overly smooth results. Concurrent works that leverage recent advances in conditional generative adversarial networks (cGANs) are proposed to enable a universal approach to diverse image translation tasks that traditionally require specific loss functions. Despite the impressive results, most of these approaches are notoriously unstable to train and tend to induce blurs. In this paper, we decompose the image into a set of images by a feature pyramid and elaborate separate loss components for images of specific bandpass. The overall perceptual adversarial loss is able to capture not only the semantic features but also the appearance
Adversarial Sparse-View CBCT Artifact Reduction
We present an effective post-processing method to reduce the artifacts from
sparsely reconstructed cone-beam CT (CBCT) images. The proposed method is based
on the state-of-the-art, image-to-image generative models with a perceptual
loss as regulation. Unlike the traditional CT artifact-reduction approaches,
our method is trained in an adversarial fashion that yields more perceptually
realistic outputs while preserving the anatomical structures. To address the
streak artifacts that are inherently local and appear across various scales, we
further propose a novel discriminator architecture based on feature pyramid
networks and a differentially modulated focus map to induce the adversarial
training. Our experimental results show that the proposed method can greatly
correct the cone-beam artifacts from clinical CBCT images reconstructed using
1/3 projections, and outperforms strong baseline methods both quantitatively
and qualitatively
High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs
We present a new method for synthesizing high-resolution photo-realistic
images from semantic label maps using conditional generative adversarial
networks (conditional GANs). Conditional GANs have enabled a variety of
applications, but the results are often limited to low-resolution and still far
from realistic. In this work, we generate 2048x1024 visually appealing results
with a novel adversarial loss, as well as new multi-scale generator and
discriminator architectures. Furthermore, we extend our framework to
interactive visual manipulation with two additional features. First, we
incorporate object instance segmentation information, which enables object
manipulations such as removing/adding objects and changing the object category.
Second, we propose a method to generate diverse results given the same input,
allowing users to edit the object appearance interactively. Human opinion
studies demonstrate that our method significantly outperforms existing methods,
advancing both the quality and the resolution of deep image synthesis and
editing.Comment: v2: CVPR camera ready, adding more results for edge-to-photo example
Adversarial nets with perceptual losses for text-to-image synthesis
Recent approaches in generative adversarial networks (GANs) can automatically
synthesize realistic images from descriptive text. Despite the overall fair
quality, the generated images often expose visible flaws that lack structural
definition for an object of interest. In this paper, we aim to extend state of
the art for GAN-based text-to-image synthesis by improving perceptual quality
of generated images. Differentiated from previous work, our synthetic image
generator optimizes on perceptual loss functions that measure pixel, feature
activation, and texture differences against a natural image. We present
visually more compelling synthetic images of birds and flowers generated from
text descriptions in comparison to some of the most prominent existing work
Contextual-based Image Inpainting: Infer, Match, and Translate
We study the task of image inpainting, which is to fill in the missing region
of an incomplete image with plausible contents. To this end, we propose a
learning-based approach to generate visually coherent completion given a
high-resolution image with missing components. In order to overcome the
difficulty to directly learn the distribution of high-dimensional image data,
we divide the task into inference and translation as two separate steps and
model each step with a deep neural network. We also use simple heuristics to
guide the propagation of local textures from the boundary to the hole. We show
that, by using such techniques, inpainting reduces to the problem of learning
two image-feature translation functions in much smaller space and hence easier
to train. We evaluate our method on several public datasets and show that we
generate results of better visual quality than previous state-of-the-art
methods.Comment: ECCV 2018 camera read
- …