126,294 research outputs found
Transferring near infrared spectroscopic calibration model across different harvested seasons using joint distribution adaptation
Near infrared spectroscopic (NIRS) data from different harvested
seasons may consist of different feature spaces even though the samples have the
same label values. This is because the spectral response could be affected by the
changes in environmental parameters, internal quality, and the reproducibility of
NIRS instruments. Thus, this study aims to investigate the ability of Joint
Distribution Adaptation (JDA) transfer learning algorithm in addressing the
assumption of traditional machine learning i.e. both training and testing data
must come from the same feature spaces and data distribution. First, NIRS data
acquired from two different harvested seasons were used as the source domain
and the target domain, respectively. Next, JDA was implemented to produce an
adaptation matrix using the source domain and transfer datasets. This adaptation
matrix would be used to transform the source and target domain datasets. After
that, a calibration model was developed by means of Partial Least Squares
(PLS) using the transformed training dataset; and validated using the trans๏ฟฝformed independent testing dataset. The proposed JDA-PLS was compared to
the PLS without transfer learning as the baseline learning. Findings show that
the proposed JDA-PLS with 10 LVs achieved the lowest RMSEP of 1.134% and
the highest RP
2 of 0.826
๊ณต๋ ๋์กฐ์ ํ์ต์ ์ด์ฉํ ๋น์ง๋ ๋๋ฉ์ธ ์ ์ ๊ธฐ๋ฒ ์ฐ๊ตฌ
ํ์๋
ผ๋ฌธ (์์ฌ) -- ์์ธ๋ํ๊ต ๋ํ์ : ๊ณต๊ณผ๋ํ ์ ๊ธฐยท์ ๋ณด๊ณตํ๋ถ, 2021. 2. ์ค์ฑ๋ก.Domain adaptation is introduced to exploit the label information of source domain when labels are not available for target domain. Previous methods minimized domain discrepancy in a latent space to enable transfer learning. These studies are based on the theoretical analysis that the target error is upper bounded by the sum of source error, the domain discrepancy, and the joint error of the ideal hypothesis. However, feature discriminability is sacrificed while enhancing the feature transferability by matching marginal distributions. In particular, the ideal joint hypothesis error in the target error upper bound, which was previously considered to be minute, has been found to be significant, impairing its theoretical guarantee.
In this paper, to manage the joint error, we propose an alternative upper bound on the target error that explicitly considers it. Based on the theoretical analysis, we suggest a joint optimization framework that combines the source and target domains. To minimize the joint error, we further introduce Joint Contrastive Learning (JCL) that finds class-level discriminative features. With a solid theoretical framework, JCL employs contrastive loss to maximize the mutual information between a feature and its label, which is equivalent to maximizing the Jensen-Shannon divergence between conditional distributions. Extensive experiments on domain adaptation datasets demonstrate that JCL outperforms existing state-of-the-art methods.๋๋ฉ์ธ ์ ์ ๊ธฐ๋ฒ์ ํ๊ฒ ๋๋ฉ์ธ์ ๋ผ๋ฒจ ์ ๋ณด๊ฐ ์๋ ์ํฉ์์ ๋น์ทํ ๋๋ฉ์ธ์ธ ์์ค ๋๋ฉ์ธ์ ๋ผ๋ฒจ ์ ๋ณด๋ฅผ ํ์ฉํ๊ธฐ ์ํด ๊ฐ๋ฐ๋์๋ค. ๊ธฐ์กด์ ๋ฐฉ๋ฒ๋ก ๋ค์ ์ ์ฌ ๊ณต๊ฐ์์ ๋๋ฉ์ธ๋ค ์ฌ์ด์ ๋ถํฌ ์ฐจ์ด๋ฅผ ์ค์์ผ๋ก์จ ์ ์ด ํ์ต์ด ๊ฐ๋ฅํ๊ฒ ํ์๋ค. ์ด๋ฌํ ๊ธฐ๋ฒ๋ค์ ์์ค ๋๋ฉ์ธ์ ์๋ฌ์จ, ๋๋ฉ์ธ ๊ฐ ๋ถํฌ ์ฐจ์ด, ๊ทธ๋ฆฌ๊ณ ์ ๋๋ฉ์ธ์์ ์ด์์ ์ธ ๋ถ๋ฅ๊ธฐ์ ์๋ฌ์จ์ ํฉ์ด ํ๊ฒ ๋๋ฉ์ธ์ ์๋ฌ์จ์ ์๊ณ๊ฐ ๋๋ค๋ ์ด๋ก ์ ๋ฐํ์ผ๋ก ํ๋ค. ๊ทธ๋ฌ๋ ๋๋ฉ์ธ๋ค ์ฌ์ด์ ๋ถํฌ ์ฐจ์ด๋ฅผ ์ค์ด๋ ๋ฐฉ๋ฒ๋ค์ ๋์์ ์ ์ฌ ๊ณต๊ฐ์์ ์๋ก ๋ค๋ฅธ ๋ผ๋ฒจ์ ๊ฐ๋ ๋ฐ์ดํฐ๋ค ์ฌ์ด์ ๊ตฌ๋ณ์ฑ์ ๊ฐ์์์ผฐ๋ค. ํนํ, ์์ ๊ฒ์ด๋ผ ์๊ฐ๋๋ ์ ๋๋ฉ์ธ์์ ์ด์์ ์ธ ๋ถ๋ฅ๊ธฐ์ ์๋ฌ์จ์ด ํฐ ๊ฒ์ผ๋ก ๋ํ๋ฌ๋ค.
๋ณธ ๋
ผ๋ฌธ์์๋ ๊ธฐ์กด์ ์ด๋ก ์์๋ ๋ค๋ฃจ์ง ์์ ์ ๋๋ฉ์ธ์์ ๋ถ๋ฅ๊ธฐ์ ์๋ฌ์จ์ ์กฐ์ ํ ์ ์๊ฒํ๊ธฐ ์ํด ์๋ก์ด ์ด๋ก ์ ์ ์ํ๋ค. ์ด ์ด๋ก ์ ๋ฐฐ๊ฒฝ์ ๋ฐํ์ผ๋ก ์์ค ๋๋ฉ์ธ๊ณผ ํ๊ฒ ๋๋ฉ์ธ์ ํจ๊ป ํ์ตํ๋ ๊ณต๋ ๋์กฐ์ ๋ฐฉ๋ฒ์ ์๊ฐํ๋ค. ๋ณธ ๊ณต๋ ๋์กฐ์ ํ์ต ๋ฐฉ๋ฒ์์๋ ๊ฐ ๋ผ๋ฒจ๋ณ๋ก ๊ตฌ๋ถ๋๋ ์ ์ฌ ๊ณต๊ฐ์ ํ์ตํ๊ธฐ ์ํด ๊ฐ ๋ฐ์ดํฐ์ ํน์ง๊ณผ ๋ผ๋ฒจ ์ฌ์ด์ ์ํธ ์ ๋ณด๋์ ์ต๋ํํ๋ค. ์ด ๊ฐ ๋ฐ์ดํฐ์ ํน์ง๊ณผ ๋ผ๋ฒจ ์ฌ์ด์ ์ํธ ์ ๋ณด๋์ ๊ฐ ๋ผ๋ฒจ ๋ถํฌ ์ฌ์ด์ ์ ์ผ-์ค๋
ผ ๊ฑฐ๋ฆฌ์ ๊ฐ์ผ๋ฏ๋ก ์ด๋ฅผ ์ต๋ํํ๋ ๊ฒ์ ๊ณง ๋ผ๋ฒจ๋ค์ด ์ ๊ตฌ๋ณ๋๋ ์ ์ฌ ๊ณต๊ฐ์ ํ์ตํ๋ ๊ฒ์ด๋ค. ๋ง์ง๋ง์ผ๋ก ๊ณต๋ ๋์กฐ์ ํ์ต ๋ฐฉ๋ฒ์ ์ฌ๋ฌ ๋ฐ์ดํฐ ์
์ ์ ์ฉํ์ฌ ๊ธฐ์กด ๋ฐฉ๋ฒ๋ก ๋ค๊ณผ ๋น๊ตํ์๋ค.1 Introduction 1
2 Background 4
2.1 Domain Adaptation . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.1.1 Problem Setting and Notations . . . . . . . . . . . . . . . . . 4
2.1.2 Theoretical Background . . . . . . . . . . . . . . . . . . . . 5
2.2 Approaches for Domain Adaptation . . . . . . . . . . . . . . . . . . 6
2.2.1 Marginal Distribution Alignment Based Approaches . . . . . 6
2.2.2 Conditional Distribution Matching Approaches . . . . . . . . 7
2.3 Contrastive Learning . . . . . . . . . . . . . . . . . . . . . . . . . . 8
3 Method 10
3.1 An Alternative Upper Bound . . . . . . . . . . . . . . . . . . . . . . 10
3.2 Joint Contrastive Learning . . . . . . . . . . . . . . . . . . . . . . . 14
3.2.1 Theoretical Guarantees . . . . . . . . . . . . . . . . . . . . . 14
3.2.2 Generalization to Multiclass Classification . . . . . . . . . . 17
3.2.3 Training Procedure . . . . . . . . . . . . . . . . . . . . . . . 19
4 Experiments 24
4.1 Datasets and Baselines . . . . . . . . . . . . . . . . . . . . . . . . . 24
4.2 Implementation Details . . . . . . . . . . . . . . . . . . . . . . . . . 26
4.3 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
4.4 Ablation Studies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
5 Conclusion 35
Abstract (In Korean) 45Maste
A review of domain adaptation without target labels
Domain adaptation has become a prominent problem setting in machine learning
and related fields. This review asks the question: how can a classifier learn
from a source domain and generalize to a target domain? We present a
categorization of approaches, divided into, what we refer to as, sample-based,
feature-based and inference-based methods. Sample-based methods focus on
weighting individual observations during training based on their importance to
the target domain. Feature-based methods revolve around on mapping, projecting
and representing features such that a source classifier performs well on the
target domain and inference-based methods incorporate adaptation into the
parameter estimation procedure, for instance through constraints on the
optimization procedure. Additionally, we review a number of conditions that
allow for formulating bounds on the cross-domain generalization error. Our
categorization highlights recurring ideas and raises questions important to
further research.Comment: 20 pages, 5 figure
Joint Distribution Optimal Transportation for Domain Adaptation
This paper deals with the unsupervised domain adaptation problem, where one
wants to estimate a prediction function in a given target domain without
any labeled sample by exploiting the knowledge available from a source domain
where labels are known. Our work makes the following assumption: there exists a
non-linear transformation between the joint feature/label space distributions
of the two domain and . We propose a solution of
this problem with optimal transport, that allows to recover an estimated target
by optimizing simultaneously the optimal coupling
and . We show that our method corresponds to the minimization of a bound on
the target error, and provide an efficient algorithmic solution, for which
convergence is proved. The versatility of our approach, both in terms of class
of hypothesis or loss functions is demonstrated with real world classification
and regression problems, for which we reach or surpass state-of-the-art
results.Comment: Accepted for publication at NIPS 201
- โฆ