5 research outputs found

    ์‹ค์‹œ๊ฐ„ ์ด๋ฏธ์ง€ ํŽธ์ง‘์„ ์œ„ํ•œ GAN ๋‚ด ์ž ์žฌ์˜ ๊ณต๊ฐ„์ฐจ์› ํ™œ์šฉ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (์„์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2021. 2. ์œ ์Šน์ฃผ.Generative adversarial networks (GANs) have been successful in synthesizing and manipulating synthetic but realistic images from latent vectors. However, it is still challenging for GANs to manipulate real images, especially in real-time. State-of-the-art GAN-based methods for editing real images suffer from time-consuming operations in projecting real images to latent vectors. Alternatively, an encoder can be trained to embed real images to the latent space instantly, but it loses details drastically. We propose StyleMapGAN, which adopts a novel representation of latent space, called stylemap, incorporating spatial dimension into embedding. Because each spatial location in the stylemap contributes to its corresponding region of the generated images, the real-time projection through the encoder becomes accurate as well as editing real images becomes spatially controllable. Experimental results demonstrate that our method significantly outperforms state-of-the-art models in various image manipulation tasks such as local editing and image interpolation. Especially, detailed comparisons show that our local editing method successfully reflects not only the color and texture but also the shape of a reference image while preserving untargeted regions.์ ๋Œ€์  ์ƒ์„ฑ ์‹ ๊ฒฝ๋ง(GAN)์€ ์‹ค์กดํ•˜์ง€ ์•Š์ง€๋งŒ, ์‹ค์ œ ์กด์žฌํ•˜๋Š” ๊ฒƒ ๊ฐ™์€ ์ด๋ฏธ์ง€๋“ค์„ ์ƒ์„ฑํ•˜๋Š”๋ฐ ์„ฑ๊ณต์ ์œผ๋กœ ์ด์šฉ๋˜๊ณ  ์žˆ๋‹ค. ๋˜ํ•œ ๊ฐ ์ด๋ฏธ์ง€๋ฅผ ์ƒ์„ฑํ•˜๋Š” ์ž ์žฌ ๋ฒกํ„ฐ๋ฅผ ์ด์šฉํ•ด ๊ฐ€์งœ(์‹ค์กดํ•˜์ง€ ์•Š๋Š”) ์ด๋ฏธ์ง€๋“ค์„ ํŽธ์ง‘ํ•  ์ˆ˜ ์žˆ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜ ๊ฐ€์งœ ์ด๋ฏธ์ง€๊ฐ€ ์•„๋‹Œ ์‹ค์ œ ์ด๋ฏธ์ง€๋ฅผ ํŽธ์ง‘ํ•˜๋Š” ๊ฒƒ์€ ์–ด๋ ต๊ณ , ํŠนํžˆ ์‹ค์‹œ๊ฐ„์œผ๋กœ๋Š” ๋”์šฑ ์–ด๋ ต๋‹ค. GAN์„ ์ด์šฉํ•ด ์‹ค์ œ ์ด๋ฏธ์ง€๋ฅผ ํŽธ์ง‘ํ•˜๋Š” ์ตœ์ฒจ๋‹จ ๋ฐฉ๋ฒ•๋“ค์€ ์‹ค์ œ ์ด๋ฏธ์ง€๋ฅผ ์ž ์žฌ ๋ฒกํ„ฐ๋กœ ํˆฌ์˜ํ•˜๋Š” ๊ฒƒ์ด ์„ ํ–‰๋˜์–ด์•ผ ํ•˜๋Š”๋ฐ, ์ด ๋ถ€๋ถ„์— ๋งŽ์€ ์‹œ๊ฐ„์ด ์†Œ์š”๋œ๋‹ค. ๊ทธ ๋Œ€์•ˆ์œผ๋กœ ์ธ์ฝ”๋”๋ฅผ ํ•™์Šตํ•ด์„œ ์‹ค์ œ ์ด๋ฏธ์ง€๋ฅผ ์ž ์žฌ ๊ณต๊ฐ„์œผ๋กœ ์ฆ‰์‹œ ์ž„๋ฒ ๋”ฉํ•  ์ˆ˜ ์žˆ์ง€๋งŒ, ์ž ์žฌ ๋ฒกํ„ฐ๋ฅผ ๋‹ค์‹œ ์ด๋ฏธ์ง€๋กœ ๋ณต์› ์‹œ ๋งŽ์€ ๋””ํ…Œ์ผ๋“ค์„ ์žƒ์–ด๋ฒ„๋ฆฐ๋‹ค. ์šฐ๋ฆฌ๋Š” ์ƒˆ๋กœ์šด ํ˜•ํƒœ์˜ ์ž ์žฌ ๊ณต๊ฐ„์ธ stylemap์„ ๊ฐ€์ง€๋Š” StyleMapGAN์„ ์ œ์•ˆํ–ˆ๋Š”๋ฐ, stylemap์€ ๊ธฐ์กด ๋ฒกํ„ฐ ๊ณต๊ฐ„์— ๊ณต๊ฐ„์ ์ธ ์ฐจ์›์„ ์ถ”๊ฐ€ํ•œ ์ž ์žฌ ๊ณต๊ฐ„์ด๋‹ค. Stylemap์˜ ๊ฐ ์œ„์น˜๋Š” ์ƒ์„ฑ๋œ ์ด๋ฏธ์ง€์˜ ํ•ด๋‹น ์ง€์—ญ์— ๋Œ€์‘๋˜์–ด, ์ธ์ฝ”๋”๋ฅผ ์ด์šฉํ•ด ์‹ค์‹œ๊ฐ„์ด๋ฉด์„œ๋„ ์ •ํ™•ํ•œ ํˆฌ์˜์ด ๊ฐ€๋Šฅํ•ด์งˆ ๋ฟ๋งŒ ์•„๋‹ˆ๋ผ, ์‹ค์ œ ์ด๋ฏธ์ง€์˜ ๊ณต๊ฐ„์ ์œผ๋กœ ํŽธ์ง‘์ด ๊ฐ€๋Šฅํ•ด์ง„๋‹ค. ๋งŽ์€ ์‹คํ—˜ ๊ฒฐ๊ณผ๋“ค์€ ์šฐ๋ฆฌ๊ฐ€ ์ œ์•ˆํ•œ ๋ฐฉ๋ฒ•์ด ๋‹ค์–‘ํ•œ ์ด๋ฏธ์ง€ ํŽธ์ง‘ ์ž‘์—…๋“ค(์˜ˆ๋ฅผ ๋“ค์–ด, ๊ตญ์†Œ์ ์ธ ํŽธ์ง‘ ๋ฐ ์ด๋ฏธ์ง€ ๋ณด๊ฐ„)์—์„œ ๊ธฐ์กด ์ตœ์ฒจ๋‹จ ๋ฐฉ๋ฒ•๋“ค์„ ์›”๋“ฑํžˆ ๋Šฅ๊ฐ€ํ•จ์„ ๋ณด์—ฌ์ค€๋‹ค. ํŠนํžˆ ์ž์„ธํ•œ ๋น„๊ต ์‹คํ—˜๋“ค์€ ์šฐ๋ฆฌ๊ฐ€ ์ œ์•ˆํ•œ ๊ตญ์†Œ์ ์ธ ํŽธ์ง‘ ๋ฐฉ๋ฒ•์ด ํšจ๊ณผ์ ์ž„์„ ๋ณด์—ฌ์ค€๋‹ค. ํƒ€๊ฒŸํ•˜์ง€ ์•Š๋Š” ๋ถ€๋ถ„์˜ ๊ธฐ์กด ์ด๋ฏธ์ง€๋Š” ์ž˜ ์œ ์ง€๋˜๊ณ , ํƒ€๊ฒŸํ•˜๋Š” ๋ถ€๋ถ„์—์„œ๋Š” ์ฐธ์กฐ ์ด๋ฏธ์ง€์˜ ์ƒ‰๊ณผ ์งˆ๊ฐ๋ฟ๋งŒ ์•„๋‹ˆ๋ผ ๋ชจ์–‘๊นŒ์ง€ ์ž˜ ๊ฐ€์ ธ์˜ด์„ ๋ณผ ์ˆ˜ ์žˆ๋‹ค.Abstract i Contents iii List of Tables v List of Figures vii Chapter 1 Introduction 1 Chapter 2 Related Work 3 Chapter 3 StyleMapGAN 5 3.1 Stylemap-based generator 7 3.2 Training procedure and losses 10 3.3 Local editing 10 Chapter 4 Experiments 12 4.1 Experimental Setup 12 4.2 Evaluation metrics 13 4.3 Effects of stylemap resolution 17 4.4 Real Image Projection 19 4.5 Local Editing 19 4.6 Unaligned Transplantation 20 Chapter 5 Discussion and Conclusion 24 References 25 Appendix 31 A Local editing in w+ space 31 B Additional results 33 B.1 Random generation 33 B.2 Image projection & Interpolation 34 B.3 Local editing 34 B.4 Unaligned transplantation 39 B.5 Style mixing 39 B.6 Semantic manipulation 39 B.7 Failure cases 40 C Implementation details 46 D Loss details 49 ๊ตญ๋ฌธ์ดˆ๋ก 51Maste
    corecore