13 research outputs found
Cross-Domain Labeled LDA for Cross-Domain Text Classification
Cross-domain text classification aims at building a classifier for a target
domain which leverages data from both source and target domain. One promising
idea is to minimize the feature distribution differences of the two domains.
Most existing studies explicitly minimize such differences by an exact
alignment mechanism (aligning features by one-to-one feature alignment,
projection matrix etc.). Such exact alignment, however, will restrict models'
learning ability and will further impair models' performance on classification
tasks when the semantic distributions of different domains are very different.
To address this problem, we propose a novel group alignment which aligns the
semantics at group level. In addition, to help the model learn better semantic
groups and semantics within these groups, we also propose a partial supervision
for model's learning in source domain. To this end, we embed the group
alignment and a partial supervision into a cross-domain topic model, and
propose a Cross-Domain Labeled LDA (CDL-LDA). On the standard 20Newsgroup and
Reuters dataset, extensive quantitative (classification, perplexity etc.) and
qualitative (topic detection) experiments are conducted to show the
effectiveness of the proposed group alignment and partial supervision.Comment: ICDM 201
Wasserstein Distance based Deep Adversarial Transfer Learning for Intelligent Fault Diagnosis
The demand of artificial intelligent adoption for condition-based maintenance
strategy is astonishingly increased over the past few years. Intelligent fault
diagnosis is one critical topic of maintenance solution for mechanical systems.
Deep learning models, such as convolutional neural networks (CNNs), have been
successfully applied to fault diagnosis tasks for mechanical systems and
achieved promising results. However, for diverse working conditions in the
industry, deep learning suffers two difficulties: one is that the well-defined
(source domain) and new (target domain) datasets are with different feature
distributions; another one is the fact that insufficient or no labelled data in
target domain significantly reduce the accuracy of fault diagnosis. As a novel
idea, deep transfer learning (DTL) is created to perform learning in the target
domain by leveraging information from the relevant source domain. Inspired by
Wasserstein distance of optimal transport, in this paper, we propose a novel
DTL approach to intelligent fault diagnosis, namely Wasserstein Distance based
Deep Transfer Learning (WD-DTL), to learn domain feature representations
(generated by a CNN based feature extractor) and to minimize the distributions
between the source and target domains through adversarial training. The
effectiveness of the proposed WD-DTL is verified through 3 transfer scenarios
and 16 transfer fault diagnosis experiments of both unsupervised and supervised
(with insufficient labelled data) learning. We also provide a comprehensive
analysis of the network visualization of those transfer tasks