46,297 research outputs found
SizeNet: Weakly Supervised Learning of Visual Size and Fit in Fashion Images
Finding clothes that fit is a hot topic in the e-commerce fashion industry.
Most approaches addressing this problem are based on statistical methods
relying on historical data of articles purchased and returned to the store.
Such approaches suffer from the cold start problem for the thousands of
articles appearing on the shopping platforms every day, for which no prior
purchase history is available. We propose to employ visual data to infer size
and fit characteristics of fashion articles. We introduce SizeNet, a
weakly-supervised teacher-student training framework that leverages the power
of statistical models combined with the rich visual information from article
images to learn visual cues for size and fit characteristics, capable of
tackling the challenging cold start problem. Detailed experiments are performed
on thousands of textile garments, including dresses, trousers, knitwear, tops,
etc. from hundreds of different brands.Comment: IEEE Conference on Computer Vision and Pattern Recognition Workshop
(CVPRW) 2019 Focus on Fashion and Subjective Search - Understanding
Subjective Attributes of Data (FFSS-USAD
Creating Capsule Wardrobes from Fashion Images
We propose to automatically create capsule wardrobes. Given an inventory of
candidate garments and accessories, the algorithm must assemble a minimal set
of items that provides maximal mix-and-match outfits. We pose the task as a
subset selection problem. To permit efficient subset selection over the space
of all outfit combinations, we develop submodular objective functions capturing
the key ingredients of visual compatibility, versatility, and user-specific
preference. Since adding garments to a capsule only expands its possible
outfits, we devise an iterative approach to allow near-optimal submodular
function maximization. Finally, we present an unsupervised approach to learn
visual compatibility from "in the wild" full body outfit photos; the
compatibility metric translates well to cleaner catalog photos and improves
over existing methods. Our results on thousands of pieces from popular fashion
websites show that automatic capsule creation has potential to mimic skilled
fashionistas in assembling flexible wardrobes, while being significantly more
scalable.Comment: Accepted to CVPR 201
Multi-modal Embedding Fusion-based Recommender
Recommendation systems have lately been popularized globally, with primary
use cases in online interaction systems, with significant focus on e-commerce
platforms. We have developed a machine learning-based recommendation platform,
which can be easily applied to almost any items and/or actions domain. Contrary
to existing recommendation systems, our platform supports multiple types of
interaction data with multiple modalities of metadata natively. This is
achieved through multi-modal fusion of various data representations. We
deployed the platform into multiple e-commerce stores of different kinds, e.g.
food and beverages, shoes, fashion items, telecom operators. Here, we present
our system, its flexibility and performance. We also show benchmark results on
open datasets, that significantly outperform state-of-the-art prior work.Comment: 7 pages, 8 figure
Image-based Recommendations on Styles and Substitutes
Humans inevitably develop a sense of the relationships between objects, some
of which are based on their appearance. Some pairs of objects might be seen as
being alternatives to each other (such as two pairs of jeans), while others may
be seen as being complementary (such as a pair of jeans and a matching shirt).
This information guides many of the choices that people make, from buying
clothes to their interactions with each other. We seek here to model this human
sense of the relationships between objects based on their appearance. Our
approach is not based on fine-grained modeling of user annotations but rather
on capturing the largest dataset possible and developing a scalable method for
uncovering human notions of the visual relationships within. We cast this as a
network inference problem defined on graphs of related images, and provide a
large-scale dataset for the training and evaluation of the same. The system we
develop is capable of recommending which clothes and accessories will go well
together (and which will not), amongst a host of other applications.Comment: 11 pages, 10 figures, SIGIR 201
- …