1,867 research outputs found
Open-world Learning and Application to Product Classification
Classic supervised learning makes the closed-world assumption, meaning that
classes seen in testing must have been seen in training. However, in the
dynamic world, new or unseen class examples may appear constantly. A model
working in such an environment must be able to reject unseen classes (not seen
or used in training). If enough data is collected for the unseen classes, the
system should incrementally learn to accept/classify them. This learning
paradigm is called open-world learning (OWL). Existing OWL methods all need
some form of re-training to accept or include the new classes in the overall
model. In this paper, we propose a meta-learning approach to the problem. Its
key novelty is that it only needs to train a meta-classifier, which can then
continually accept new classes when they have enough labeled data for the
meta-classifier to use, and also detect/reject future unseen classes. No
re-training of the meta-classifier or a new overall classifier covering all old
and new classes is needed. In testing, the method only uses the examples of the
seen classes (including the newly added classes) on-the-fly for classification
and rejection. Experimental results demonstrate the effectiveness of the new
approach.Comment: accepted by The Web Conference (WWW 2019) Previous title: Learning to
Accept New Classes without Trainin
Learning and Transferring IDs Representation in E-commerce
Many machine intelligence techniques are developed in E-commerce and one of
the most essential components is the representation of IDs, including user ID,
item ID, product ID, store ID, brand ID, category ID etc. The classical
encoding based methods (like one-hot encoding) are inefficient in that it
suffers sparsity problems due to its high dimension, and it cannot reflect the
relationships among IDs, either homogeneous or heterogeneous ones. In this
paper, we propose an embedding based framework to learn and transfer the
representation of IDs. As the implicit feedbacks of users, a tremendous amount
of item ID sequences can be easily collected from the interactive sessions. By
jointly using these informative sequences and the structural connections among
IDs, all types of IDs can be embedded into one low-dimensional semantic space.
Subsequently, the learned representations are utilized and transferred in four
scenarios: (i) measuring the similarity between items, (ii) transferring from
seen items to unseen items, (iii) transferring across different domains, (iv)
transferring across different tasks. We deploy and evaluate the proposed
approach in Hema App and the results validate its effectiveness.Comment: KDD'18, 9 page
- …