31 research outputs found
Unsupervised Domain Adaptation by Backpropagation
Top-performing deep architectures are trained on massive amounts of labeled
data. In the absence of labeled data for a certain task, domain adaptation
often provides an attractive option given that labeled data of similar nature
but from a different domain (e.g. synthetic images) are available. Here, we
propose a new approach to domain adaptation in deep architectures that can be
trained on large amount of labeled data from the source domain and large amount
of unlabeled data from the target domain (no labeled target-domain data is
necessary).
As the training progresses, the approach promotes the emergence of "deep"
features that are (i) discriminative for the main learning task on the source
domain and (ii) invariant with respect to the shift between the domains. We
show that this adaptation behaviour can be achieved in almost any feed-forward
model by augmenting it with few standard layers and a simple new gradient
reversal layer. The resulting augmented architecture can be trained using
standard backpropagation.
Overall, the approach can be implemented with little effort using any of the
deep-learning packages. The method performs very well in a series of image
classification experiments, achieving adaptation effect in the presence of big
domain shifts and outperforming previous state-of-the-art on Office datasets
Domain-Adversarial Training of Neural Networks
We introduce a new representation learning approach for domain adaptation, in
which data at training and test time come from similar but different
distributions. Our approach is directly inspired by the theory on domain
adaptation suggesting that, for effective domain transfer to be achieved,
predictions must be made based on features that cannot discriminate between the
training (source) and test (target) domains. The approach implements this idea
in the context of neural network architectures that are trained on labeled data
from the source domain and unlabeled data from the target domain (no labeled
target-domain data is necessary). As the training progresses, the approach
promotes the emergence of features that are (i) discriminative for the main
learning task on the source domain and (ii) indiscriminate with respect to the
shift between the domains. We show that this adaptation behaviour can be
achieved in almost any feed-forward model by augmenting it with few standard
layers and a new gradient reversal layer. The resulting augmented architecture
can be trained using standard backpropagation and stochastic gradient descent,
and can thus be implemented with little effort using any of the deep learning
packages. We demonstrate the success of our approach for two distinct
classification problems (document sentiment analysis and image classification),
where state-of-the-art domain adaptation performance on standard benchmarks is
achieved. We also validate the approach for descriptor learning task in the
context of person re-identification application.Comment: Published in JMLR: http://jmlr.org/papers/v17/15-239.htm
Reimagining City Configuration: Automated Urban Planning via Adversarial Learning
Urban planning refers to the efforts of designing land-use configurations.
Effective urban planning can help to mitigate the operational and social
vulnerability of a urban system, such as high tax, crimes, traffic congestion
and accidents, pollution, depression, and anxiety. Due to the high complexity
of urban systems, such tasks are mostly completed by professional planners.
But, human planners take longer time. The recent advance of deep learning
motivates us to ask: can machines learn at a human capability to automatically
and quickly calculate land-use configuration, so human planners can finally
adjust machine-generated plans for specific needs? To this end, we formulate
the automated urban planning problem into a task of learning to configure
land-uses, given the surrounding spatial contexts. To set up the task, we
define a land-use configuration as a longitude-latitude-channel tensor, where
each channel is a category of POIs and the value of an entry is the number of
POIs. The objective is then to propose an adversarial learning framework that
can automatically generate such tensor for an unplanned area. In particular, we
first characterize the contexts of surrounding areas of an unplanned area by
learning representations from spatial graphs using geographic and human
mobility data. Second, we combine each unplanned area and its surrounding
context representation as a tuple, and categorize all the tuples into positive
(well-planned areas) and negative samples (poorly-planned areas). Third, we
develop an adversarial land-use configuration approach, where the surrounding
context representation is fed into a generator to generate a land-use
configuration, and a discriminator learns to distinguish among positive and
negative samples.Comment: Proceedings of the 28th International Conference on Advances in
Geographic Information Systems (2020
Adversarial Adaptation of Scene Graph Models for Understanding Civic Issues
Citizen engagement and technology usage are two emerging trends driven by
smart city initiatives. Governments around the world are adopting technology
for faster resolution of civic issues. Typically, citizens report issues, such
as broken roads, garbage dumps, etc. through web portals and mobile apps, in
order for the government authorities to take appropriate actions. Several
mediums -- text, image, audio, video -- are used to report these issues.
Through a user study with 13 citizens and 3 authorities, we found that image is
the most preferred medium to report civic issues. However, analyzing civic
issue related images is challenging for the authorities as it requires manual
effort. Moreover, previous works have been limited to identifying a specific
set of issues from images. In this work, given an image, we propose to generate
a Civic Issue Graph consisting of a set of objects and the semantic relations
between them, which are representative of the underlying civic issue. We also
release two multi-modal (text and images) datasets, that can help in further
analysis of civic issues from images. We present a novel approach for
adversarial training of existing scene graph models that enables the use of
scene graphs for new applications in the absence of any labelled training data.
We conduct several experiments to analyze the efficacy of our approach, and
using human evaluation, we establish the appropriateness of our model at
representing different civic issues.Comment: Accepted at WWW'1
SAR: Learning Cross-Language API Mappings with Little Knowledge
To save effort, developers often translate programs from one programming language to another, instead of implementing it from scratch. Translating application program interfaces (APIs) used in one language to functionally equivalent ones available in another language is an important aspect of program translation. Existing approaches facilitate the translation by automatically identifying the API mappings across programming languages. However, these approaches still require a large number of parallel corpora, ranging from pairs of APIs or code fragments that are functionally equivalent, to similar code comments.
To minimize the need for parallel corpora, this paper aims at an automated approach that can map APIs across languages with much less a priori knowledge than other approaches. Our approach is based on a realization of the notion of domain adaption, combined with code embedding, to better align two vector spaces. Taking as input large sets of programs, our approach first generates numeric vector representations of the programs (including the APIs used in each language), and it adapts generative adversarial networks (GAN) to align the vectors in different spaces of two languages. For better alignment, we initialize the GAN with parameters derived from API mapping seeds that can be identified accurately with a simple automatic signature-based matching heuristic. Then the cross-language API mappings can be identified via nearest-neighbors queries in the aligned vector spaces. We have implemented the approach (SAR, named after three main technical components, Seeding, Adversarial training, and Refinement) in a prototype for mapping APIs across Java and C# programs. Our evaluation on about 2 million Java files and 1 million C# files shows that the approach can achieve 48% and 78% mapping accuracy in its top-1 and top-10 API mapping results respectively, with only 174 automatically identified seeds, which is more accurate than other approaches using the same or much more mapping seeds
Self-conditioned Embedding Diffusion for Text Generation
Can continuous diffusion models bring the same performance breakthrough on
natural language they did for image generation? To circumvent the discrete
nature of text data, we can simply project tokens in a continuous space of
embeddings, as is standard in language modeling. We propose Self-conditioned
Embedding Diffusion, a continuous diffusion mechanism that operates on token
embeddings and allows to learn flexible and scalable diffusion models for both
conditional and unconditional text generation. Through qualitative and
quantitative evaluation, we show that our text diffusion models generate
samples comparable with those produced by standard autoregressive language
models - while being in theory more efficient on accelerator hardware at
inference time. Our work paves the way for scaling up diffusion models for
text, similarly to autoregressive models, and for improving performance with
recent refinements to continuous diffusion.Comment: 15 page