196 research outputs found
Improving Sketch Colorization using Adversarial Segmentation Consistency
We propose a new method for producing color images from sketches. Current
solutions in sketch colorization either necessitate additional user instruction
or are restricted to the "paired" translation strategy. We leverage semantic
image segmentation from a general-purpose panoptic segmentation network to
generate an additional adversarial loss function. The proposed loss function is
compatible with any GAN model. Our method is not restricted to datasets with
segmentation labels and can be applied to unpaired translation tasks as well.
Using qualitative, and quantitative analysis, and based on a user study, we
demonstrate the efficacy of our method on four distinct image datasets. On the
FID metric, our model improves the baseline by up to 35 points. Our code,
pretrained models, scripts to produce newly introduced datasets and
corresponding sketch images are available at
https://github.com/giddyyupp/AdvSegLoss.Comment: Under review at Pattern Recognition Letters. arXiv admin note:
substantial text overlap with arXiv:2102.0619
A survey of comics research in computer science
Graphical novels such as comics and mangas are well known all over the world.
The digital transition started to change the way people are reading comics,
more and more on smartphones and tablets and less and less on paper. In the
recent years, a wide variety of research about comics has been proposed and
might change the way comics are created, distributed and read in future years.
Early work focuses on low level document image analysis: indeed comic books are
complex, they contains text, drawings, balloon, panels, onomatopoeia, etc.
Different fields of computer science covered research about user interaction
and content generation such as multimedia, artificial intelligence,
human-computer interaction, etc. with different sets of values. We propose in
this paper to review the previous research about comics in computer science, to
state what have been done and to give some insights about the main outlooks
TextureGAN: Controlling Deep Image Synthesis with Texture Patches
In this paper, we investigate deep image synthesis guided by sketch, color,
and texture. Previous image synthesis methods can be controlled by sketch and
color strokes but we are the first to examine texture control. We allow a user
to place a texture patch on a sketch at arbitrary locations and scales to
control the desired output texture. Our generative network learns to synthesize
objects consistent with these texture suggestions. To achieve this, we develop
a local texture loss in addition to adversarial and content loss to train the
generative network. We conduct experiments using sketches generated from real
images and textures sampled from a separate texture database and results show
that our proposed algorithm is able to generate plausible images that are
faithful to user controls. Ablation studies show that our proposed pipeline can
generate more realistic images than adapting existing methods directly.Comment: CVPR 2018 spotligh
Image-to-Image Translation with Conditional Adversarial Networks
We investigate conditional adversarial networks as a general-purpose solution
to image-to-image translation problems. These networks not only learn the
mapping from input image to output image, but also learn a loss function to
train this mapping. This makes it possible to apply the same generic approach
to problems that traditionally would require very different loss formulations.
We demonstrate that this approach is effective at synthesizing photos from
label maps, reconstructing objects from edge maps, and colorizing images, among
other tasks. Indeed, since the release of the pix2pix software associated with
this paper, a large number of internet users (many of them artists) have posted
their own experiments with our system, further demonstrating its wide
applicability and ease of adoption without the need for parameter tweaking. As
a community, we no longer hand-engineer our mapping functions, and this work
suggests we can achieve reasonable results without hand-engineering our loss
functions either.Comment: Website: https://phillipi.github.io/pix2pix/, CVPR 201
Semantic Photo Manipulation with a Generative Image Prior
Despite the recent success of GANs in synthesizing images conditioned on
inputs such as a user sketch, text, or semantic labels, manipulating the
high-level attributes of an existing natural photograph with GANs is
challenging for two reasons. First, it is hard for GANs to precisely reproduce
an input image. Second, after manipulation, the newly synthesized pixels often
do not fit the original image. In this paper, we address these issues by
adapting the image prior learned by GANs to image statistics of an individual
image. Our method can accurately reconstruct the input image and synthesize new
content, consistent with the appearance of the input image. We demonstrate our
interactive system on several semantic image editing tasks, including
synthesizing new objects consistent with background, removing unwanted objects,
and changing the appearance of an object. Quantitative and qualitative
comparisons against several existing methods demonstrate the effectiveness of
our method.Comment: SIGGRAPH 201
๋ณํ๋ FusionNet์ ์ด์ฉํ ํ์์กฐ ์ด๋ฏธ์ง์ ์์ฐ์ค๋ฌ์ด ์ฑ์
ํ์๋
ผ๋ฌธ (์์ฌ) -- ์์ธ๋ํ๊ต ๋ํ์ : ์์ฐ๊ณผํ๋ํ ํ๋๊ณผ์ ๊ณ์ฐ๊ณผํ์ ๊ณต, 2021. 2. ๊ฐ๋ช
์ฃผ.In this paper, we propose a grayscale image colorizing technique. The colorization task can be divided into three main ways, the Scribble-based method, Exemplar-based method and Fully automatic method. Our proposed method is included in the third one. We use a deep learning model that is widely used in the colorization eld recently. We propose Encoder-Docoder model using Convolutional Neural Networks. In particular, we modify the FusionNet with good performance to suit this purpose.
Also, in order to get better results, we do not use MSE loss function. Instead, we use the loss function suitable for the colorizing purpose. We use a subset of the ImageNet dataset as the training, validation and test dataset. We take some existing methods from Fully automatic Deep Learning method and compared them with our models. Our algorithm is evaluated using a quantitative metric called PSNR (Peak Signal-to-Noise Ratio). In addition, in order to evaluate the results qualitatively, our model was applied to the test dataset and compared with various other models. Our model has better performance both quantitatively and qualitatively than other models. Finally, we apply our model to old black and white photographs.๋ณธ ๋
ผ๋ฌธ์์๋ ํ์์กฐ ์ด๋ฏธ์ง๋ค์ ๋ํ ์ฑ์ ๊ธฐ๋ฒ์ ์ ์ํ๋ค. ์ฑ์ ์์
์ ํฌ๊ฒ Scribble ๊ธฐ๋ฐ ๋ฐฉ๋ฒ, Exemplar ๊ธฐ๋ฐ ๋ฐฉ๋ฒ, ์์ ์๋ ๋ฐฉ๋ฒ์ ์ธ ๊ฐ์ง๋ก ๋๋ ์ ์๋ค. ๋ณธ ๋
ผ๋ฌธ์์๋ ์ธ ๋ฒ์งธ ๋ฐฉ๋ฒ์ ์ฌ์ฉํ๋ค. ์ต๊ทผ์ ์ฑ์ ๋ถ์ผ์์ ๋๋ฆฌ ์ฌ์ฉ๋๋ ๋ฅ ๋ฌ๋ ๋ชจ๋ธ์ ์ฌ์ฉํ๋ค. Convolutional Neural Networks๋ฅผ ์ด์ฉํ Encoder-Docoder ๋ชจ๋ธ์ ์ ์ํ๋ค. ํนํ ๊ธฐ์กด์ image segmetation ๋ถ์ผ์์ ์ข์ ์ฑ๋ฅ์ ๋ณด์ด๋ FusionNet์ ์๋ ์ฑ์ ๋ชฉ์ ์ ๋ง๊ฒ ๋ค์ํ ๋ฐฉ๋ฒ์ผ๋ก ์์ ํ๋ค. ๋ํ ๋ ๋์ ๊ฒฐ๊ณผ๋ฅผ ์ป๊ธฐ ์ํด MSE ์์ค ํจ์๋ฅผ ์ฌ์ฉํ์ง ์์๋ค. ๋์ , ์ฐ๋ฆฌ๋ ์๋ ์ฑ์ ๋ชฉ์ ์ ์ ํฉํ ์์ค ํจ์๋ฅผ ์ฌ์ฉํ์๋ค.
ImageNet ๋ฐ์ดํฐ์
์ ๋ถ๋ถ ์งํฉ์ ํ๋ จ, ๊ฒ์ฆ ๋ฐ ํ
์คํธ ๋ฐ์ดํฐ์
์ผ๋ก ์ฌ์ฉํ๋ค. ์ฐ๋ฆฌ๋ ์์ ์๋ ๋ฅ ๋ฌ๋ ๋ฐฉ๋ฒ์์ ๊ธฐ์กด ๋ฐฉ๋ฒ์ ๊ฐ์ ธ์ ์ฐ๋ฆฌ์ ๋ชจ๋ธ๊ณผ ๋น๊ตํ๋ค. ์ฐ๋ฆฌ์ ์๊ณ ๋ฆฌ์ฆ์ PSNR (Peak Signal-to-Noise Ratio)์ด๋ผ๋ ์ ๋์ ์งํ๋ฅผ ์ฌ์ฉํ์ฌ ํ๊ฐ๋์๋ค. ๋ํ ๊ฒฐ๊ณผ๋ฅผ ์ ์ฑ์ ์ผ๋ก ํ๊ฐํ๊ธฐ ์ํด ํ
์คํธ ๋ฐ์ดํฐ์
์ ๋ชจ๋ธ์ ์ ์ฉํ์ฌ ๋ค์ํ ๋ชจ๋ธ๊ณผ ๋น๊ตํ๋ค. ๊ทธ ๊ฒฐ๊ณผ ๋ค๋ฅธ ๋ชจ๋ธ์ ๋นํด ์ ์ฑ์ ์ผ๋ก๋, ์ ๋์ ์ผ๋ก๋ ์ข์ ์ฑ๋ฅ์ ๋ณด์๋ค. ๋ง์ง๋ง์ผ๋ก ์ค๋๋ ํ๋ฐฑ ์ฌ์ง๊ณผ ๊ฐ์ ๋ค์ํ ์ ํ์ ์ด๋ฏธ์ง์ ์ ์ฉํ ๊ฒฐ๊ณผ๋ฅผ ์ ์ํ๋ค.Abstract i
1 Introduction 1
2 Related Works 4
2.1 Scribble-based method . . . . . . . . . . . . . . . . . . . . . . 4
2.2 Exemplar-based method . . . . . . . . . . . . . . . . . . . . . 5
2.3 Fully automatic method . . . . . . . . . . . . . . . . . . . . . 6
3 Proposed Method 8
3.1 Method Overview . . . . . . . . . . . . . . . . . . . . . . . . . 8
3.2 Loss Function . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
3.3 Architecture detail . . . . . . . . . . . . . . . . . . . . . . . . 10
3.3.1 Encoder . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3.3.2 Decoder . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3.3.3 Bridge . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
4 Experiments 14
4.1 CIE Lab Color Space . . . . . . . . . . . . . . . . . . . . . . . 15
4.2 Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
4.3 Qualitative Evaluation . . . . . . . . . . . . . . . . . . . . . . 17
4.4 Quantitative Evaluation . . . . . . . . . . . . . . . . . . . . . 18
4.5 Legacy Old image Colorization . . . . . . . . . . . . . . . . . . 20
5 Conclusion 23
The bibliography 24
Abstract (in Korean) 28Maste
A review of image and video colorization: From analogies to deep learning
Image colorization is a classic and important topic in computer graphics, where the aim is to add color to a monochromatic input image to produce a colorful result. In this survey, we present the history of colorization research in chronological order and summarize popular algorithms in this field. Early works on colorization mostly focused on developing techniques to improve the colorization quality. In the last few years, researchers have considered more possibilities such as combining colorization with NLP (natural language processing) and focused more on industrial applications. To better control the color, various types of color control are designed, such as providing reference images or color-scribbles. We have created a taxonomy of the colorization methods according to the input type, divided into grayscale, sketch-based and hybrid. The pros and cons are discussed for each algorithm, and they are compared according to their main characteristics. Finally, we discuss how deep learning, and in particular Generative Adversarial Networks (GANs), has changed this field
- โฆ