126,294 research outputs found

    Transferring near infrared spectroscopic calibration model across different harvested seasons using joint distribution adaptation

    Get PDF
    Near infrared spectroscopic (NIRS) data from different harvested seasons may consist of different feature spaces even though the samples have the same label values. This is because the spectral response could be affected by the changes in environmental parameters, internal quality, and the reproducibility of NIRS instruments. Thus, this study aims to investigate the ability of Joint Distribution Adaptation (JDA) transfer learning algorithm in addressing the assumption of traditional machine learning i.e. both training and testing data must come from the same feature spaces and data distribution. First, NIRS data acquired from two different harvested seasons were used as the source domain and the target domain, respectively. Next, JDA was implemented to produce an adaptation matrix using the source domain and transfer datasets. This adaptation matrix would be used to transform the source and target domain datasets. After that, a calibration model was developed by means of Partial Least Squares (PLS) using the transformed training dataset; and validated using the trans๏ฟฝformed independent testing dataset. The proposed JDA-PLS was compared to the PLS without transfer learning as the baseline learning. Findings show that the proposed JDA-PLS with 10 LVs achieved the lowest RMSEP of 1.134% and the highest RP 2 of 0.826

    ๊ณต๋™ ๋Œ€์กฐ์  ํ•™์Šต์„ ์ด์šฉํ•œ ๋น„์ง€๋„ ๋„๋ฉ”์ธ ์ ์‘ ๊ธฐ๋ฒ• ์—ฐ๊ตฌ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (์„์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์ „๊ธฐยท์ •๋ณด๊ณตํ•™๋ถ€, 2021. 2. ์œค์„ฑ๋กœ.Domain adaptation is introduced to exploit the label information of source domain when labels are not available for target domain. Previous methods minimized domain discrepancy in a latent space to enable transfer learning. These studies are based on the theoretical analysis that the target error is upper bounded by the sum of source error, the domain discrepancy, and the joint error of the ideal hypothesis. However, feature discriminability is sacrificed while enhancing the feature transferability by matching marginal distributions. In particular, the ideal joint hypothesis error in the target error upper bound, which was previously considered to be minute, has been found to be significant, impairing its theoretical guarantee. In this paper, to manage the joint error, we propose an alternative upper bound on the target error that explicitly considers it. Based on the theoretical analysis, we suggest a joint optimization framework that combines the source and target domains. To minimize the joint error, we further introduce Joint Contrastive Learning (JCL) that finds class-level discriminative features. With a solid theoretical framework, JCL employs contrastive loss to maximize the mutual information between a feature and its label, which is equivalent to maximizing the Jensen-Shannon divergence between conditional distributions. Extensive experiments on domain adaptation datasets demonstrate that JCL outperforms existing state-of-the-art methods.๋„๋ฉ”์ธ ์ ์‘ ๊ธฐ๋ฒ•์€ ํƒ€๊ฒŸ ๋„๋ฉ”์ธ์˜ ๋ผ๋ฒจ ์ •๋ณด๊ฐ€ ์—†๋Š” ์ƒํ™ฉ์—์„œ ๋น„์Šทํ•œ ๋„๋ฉ”์ธ์ธ ์†Œ์Šค ๋„๋ฉ”์ธ์˜ ๋ผ๋ฒจ ์ •๋ณด๋ฅผ ํ™œ์šฉํ•˜๊ธฐ ์œ„ํ•ด ๊ฐœ๋ฐœ๋˜์—ˆ๋‹ค. ๊ธฐ์กด์˜ ๋ฐฉ๋ฒ•๋ก ๋“ค์€ ์ž ์žฌ ๊ณต๊ฐ„์—์„œ ๋„๋ฉ”์ธ๋“ค ์‚ฌ์ด์˜ ๋ถ„ํฌ ์ฐจ์ด๋ฅผ ์ค„์ž„์œผ๋กœ์จ ์ „์ด ํ•™์Šต์ด ๊ฐ€๋Šฅํ•˜๊ฒŒ ํ•˜์˜€๋‹ค. ์ด๋Ÿฌํ•œ ๊ธฐ๋ฒ•๋“ค์€ ์†Œ์Šค ๋„๋ฉ”์ธ์˜ ์—๋Ÿฌ์œจ, ๋„๋ฉ”์ธ ๊ฐ„ ๋ถ„ํฌ ์ฐจ์ด, ๊ทธ๋ฆฌ๊ณ  ์–‘ ๋„๋ฉ”์ธ์—์„œ ์ด์ƒ์ ์ธ ๋ถ„๋ฅ˜๊ธฐ์˜ ์—๋Ÿฌ์œจ์˜ ํ•ฉ์ด ํƒ€๊ฒŸ ๋„๋ฉ”์ธ์˜ ์—๋Ÿฌ์œจ์˜ ์ƒ๊ณ„๊ฐ€ ๋œ๋‹ค๋Š” ์ด๋ก ์„ ๋ฐ”ํƒ•์œผ๋กœ ํ•œ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜ ๋„๋ฉ”์ธ๋“ค ์‚ฌ์ด์˜ ๋ถ„ํฌ ์ฐจ์ด๋ฅผ ์ค„์ด๋Š” ๋ฐฉ๋ฒ•๋“ค์€ ๋™์‹œ์— ์ž ์žฌ ๊ณต๊ฐ„์—์„œ ์„œ๋กœ ๋‹ค๋ฅธ ๋ผ๋ฒจ์„ ๊ฐ–๋Š” ๋ฐ์ดํ„ฐ๋“ค ์‚ฌ์ด์˜ ๊ตฌ๋ณ„์„ฑ์„ ๊ฐ์†Œ์‹œ์ผฐ๋‹ค. ํŠนํžˆ, ์ž‘์„ ๊ฒƒ์ด๋ผ ์ƒ๊ฐ๋˜๋˜ ์–‘ ๋„๋ฉ”์ธ์—์„œ ์ด์ƒ์ ์ธ ๋ถ„๋ฅ˜๊ธฐ์˜ ์—๋Ÿฌ์œจ์ด ํฐ ๊ฒƒ์œผ๋กœ ๋‚˜ํƒ€๋‚ฌ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ๊ธฐ์กด์˜ ์ด๋ก ์—์„œ๋Š” ๋‹ค๋ฃจ์ง€ ์•Š์€ ์–‘ ๋„๋ฉ”์ธ์—์„œ ๋ถ„๋ฅ˜๊ธฐ์˜ ์—๋Ÿฌ์œจ์„ ์กฐ์ ˆํ•  ์ˆ˜ ์žˆ๊ฒŒํ•˜๊ธฐ ์œ„ํ•ด ์ƒˆ๋กœ์šด ์ด๋ก ์„ ์ œ์‹œํ•œ๋‹ค. ์ด ์ด๋ก ์  ๋ฐฐ๊ฒฝ์„ ๋ฐ”ํƒ•์œผ๋กœ ์†Œ์Šค ๋„๋ฉ”์ธ๊ณผ ํƒ€๊ฒŸ ๋„๋ฉ”์ธ์„ ํ•จ๊ป˜ ํ•™์Šตํ•˜๋Š” ๊ณต๋™ ๋Œ€์กฐ์  ๋ฐฉ๋ฒ•์„ ์†Œ๊ฐœํ•œ๋‹ค. ๋ณธ ๊ณต๋™ ๋Œ€์กฐ์  ํ•™์Šต ๋ฐฉ๋ฒ•์—์„œ๋Š” ๊ฐ ๋ผ๋ฒจ๋ณ„๋กœ ๊ตฌ๋ถ„๋˜๋Š” ์ž ์žฌ ๊ณต๊ฐ„์„ ํ•™์Šตํ•˜๊ธฐ ์œ„ํ•ด ๊ฐ ๋ฐ์ดํ„ฐ์˜ ํŠน์ง•๊ณผ ๋ผ๋ฒจ ์‚ฌ์ด์˜ ์ƒํ˜ธ ์ •๋ณด๋Ÿ‰์„ ์ตœ๋Œ€ํ™”ํ•œ๋‹ค. ์ด ๊ฐ ๋ฐ์ดํ„ฐ์˜ ํŠน์ง•๊ณผ ๋ผ๋ฒจ ์‚ฌ์ด์˜ ์ƒํ˜ธ ์ •๋ณด๋Ÿ‰์€ ๊ฐ ๋ผ๋ฒจ ๋ถ„ํฌ ์‚ฌ์ด์˜ ์  ์„ผ-์ƒค๋…ผ ๊ฑฐ๋ฆฌ์™€ ๊ฐ™์œผ๋ฏ€๋กœ ์ด๋ฅผ ์ตœ๋Œ€ํ™”ํ•˜๋Š” ๊ฒƒ์€ ๊ณง ๋ผ๋ฒจ๋“ค์ด ์ž˜ ๊ตฌ๋ณ„๋˜๋Š” ์ž ์žฌ ๊ณต๊ฐ„์„ ํ•™์Šตํ•˜๋Š” ๊ฒƒ์ด๋‹ค. ๋งˆ์ง€๋ง‰์œผ๋กœ ๊ณต๋™ ๋Œ€์กฐ์  ํ•™์Šต ๋ฐฉ๋ฒ•์„ ์—ฌ๋Ÿฌ ๋ฐ์ดํ„ฐ ์…‹์— ์ ์šฉํ•˜์—ฌ ๊ธฐ์กด ๋ฐฉ๋ฒ•๋ก ๋“ค๊ณผ ๋น„๊ตํ•˜์˜€๋‹ค.1 Introduction 1 2 Background 4 2.1 Domain Adaptation . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 2.1.1 Problem Setting and Notations . . . . . . . . . . . . . . . . . 4 2.1.2 Theoretical Background . . . . . . . . . . . . . . . . . . . . 5 2.2 Approaches for Domain Adaptation . . . . . . . . . . . . . . . . . . 6 2.2.1 Marginal Distribution Alignment Based Approaches . . . . . 6 2.2.2 Conditional Distribution Matching Approaches . . . . . . . . 7 2.3 Contrastive Learning . . . . . . . . . . . . . . . . . . . . . . . . . . 8 3 Method 10 3.1 An Alternative Upper Bound . . . . . . . . . . . . . . . . . . . . . . 10 3.2 Joint Contrastive Learning . . . . . . . . . . . . . . . . . . . . . . . 14 3.2.1 Theoretical Guarantees . . . . . . . . . . . . . . . . . . . . . 14 3.2.2 Generalization to Multiclass Classification . . . . . . . . . . 17 3.2.3 Training Procedure . . . . . . . . . . . . . . . . . . . . . . . 19 4 Experiments 24 4.1 Datasets and Baselines . . . . . . . . . . . . . . . . . . . . . . . . . 24 4.2 Implementation Details . . . . . . . . . . . . . . . . . . . . . . . . . 26 4.3 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 4.4 Ablation Studies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 5 Conclusion 35 Abstract (In Korean) 45Maste

    A review of domain adaptation without target labels

    Full text link
    Domain adaptation has become a prominent problem setting in machine learning and related fields. This review asks the question: how can a classifier learn from a source domain and generalize to a target domain? We present a categorization of approaches, divided into, what we refer to as, sample-based, feature-based and inference-based methods. Sample-based methods focus on weighting individual observations during training based on their importance to the target domain. Feature-based methods revolve around on mapping, projecting and representing features such that a source classifier performs well on the target domain and inference-based methods incorporate adaptation into the parameter estimation procedure, for instance through constraints on the optimization procedure. Additionally, we review a number of conditions that allow for formulating bounds on the cross-domain generalization error. Our categorization highlights recurring ideas and raises questions important to further research.Comment: 20 pages, 5 figure

    Joint Distribution Optimal Transportation for Domain Adaptation

    Full text link
    This paper deals with the unsupervised domain adaptation problem, where one wants to estimate a prediction function ff in a given target domain without any labeled sample by exploiting the knowledge available from a source domain where labels are known. Our work makes the following assumption: there exists a non-linear transformation between the joint feature/label space distributions of the two domain Ps\mathcal{P}_s and Pt\mathcal{P}_t. We propose a solution of this problem with optimal transport, that allows to recover an estimated target Ptf=(X,f(X))\mathcal{P}^f_t=(X,f(X)) by optimizing simultaneously the optimal coupling and ff. We show that our method corresponds to the minimization of a bound on the target error, and provide an efficient algorithmic solution, for which convergence is proved. The versatility of our approach, both in terms of class of hypothesis or loss functions is demonstrated with real world classification and regression problems, for which we reach or surpass state-of-the-art results.Comment: Accepted for publication at NIPS 201
    • โ€ฆ
    corecore