CORE
🇺🇦
make metadata, not war
Services
Services overview
Explore all CORE services
Access to raw data
API
Dataset
FastSync
Content discovery
Recommender
Discovery
OAI identifiers
OAI Resolver
Managing content
Dashboard
Bespoke contracts
Consultancy services
Support us
Support us
Membership
Sponsorship
Community governance
Advisory Board
Board of supporters
Research network
About
About us
Our mission
Team
Blog
FAQs
Contact us
research
Unsupervised person image synthesis in arbitrary poses
Authors
Antonio Agudo Martínez
Francesc Moreno-Noguer
Albert Pumarola Peris
Alberto Sanfeliu Cortés
Publication date
1 January 2018
Publisher
'Institute of Electrical and Electronics Engineers (IEEE)'
Doi
Cite
Abstract
© 2018 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting /republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other worksWe present a novel approach for synthesizing photo-realistic images of people in arbitrary poses using generative adversarial learning. Given an input image of a person and a desired pose represented by a 2D skeleton, our model renders the image of the same person under the new pose, synthesizing novel views of the parts visible in the input image and hallucinating those that are not seen. This problem has recently been addressed in a supervised manner, i.e., during training the ground truth images under the new poses are given to the network. We go beyond these approaches by proposing a fully unsupervised strategy. We tackle this challenging scenario by splitting the problem into two principal subtasks. First, we consider a pose conditioned bidirectional generator that maps back the initially rendered image to the original pose, hence being directly comparable to the input image without the need to resort to any training image. Second, we devise a novel loss function that incorporates content and style terms, and aims at producing images of high perceptual quality. Extensive experiments conducted on the DeepFashion dataset demonstrate that the images rendered by our model are very close in appearance to those obtained by fully supervised approaches.Peer ReviewedPostprint (author's final draft
Similar works
Full text
Open in the Core reader
Download PDF
Available Versions
Digital.CSIC
See this paper in CORE
Go to the repository landing page
Download from data provider
oai:digital.csic.es:10261/1794...
Last time updated on 15/05/2019
UPCommons. Portal del coneixement obert de la UPC
See this paper in CORE
Go to the repository landing page
Download from data provider
oai:upcommons.upc.edu:2117/129...
Last time updated on 06/03/2019
UPCommons
See this paper in CORE
Go to the repository landing page
Download from data provider
oai:upcommons.upc.edu:2117/129...
Last time updated on 17/04/2020
Crossref
See this paper in CORE
Go to the repository landing page
Download from data provider
Last time updated on 10/08/2021