403 research outputs found

    Deep Supervised Hashing using Symmetric Relative Entropy

    Get PDF
    By virtue of their simplicity and efficiency, hashing algorithms have achieved significant success on large-scale approximate nearest neighbor search. Recently, many deep neural network based hashing methods have been proposed to improve the search accuracy by simultaneously learning both the feature representation and the binary hash functions. Most deep hashing methods depend on supervised semantic label information for preserving the distance or similarity between local structures, which unfortunately ignores the global distribution of the learned hash codes. We propose a novel deep supervised hashing method that aims to minimize the information loss generated during the embedding process. Specifically, the information loss is measured by the Jensen-Shannon divergence to ensure that compact hash codes have a similar distribution with those from the original images. Experimental results show that our method outperforms current state-of-the-art approaches on two benchmark datasets

    Hashing for Similarity Search: A Survey

    Full text link
    Similarity search (nearest neighbor search) is a problem of pursuing the data items whose distances to a query item are the smallest from a large database. Various methods have been developed to address this problem, and recently a lot of efforts have been devoted to approximate search. In this paper, we present a survey on one of the main solutions, hashing, which has been widely studied since the pioneering work locality sensitive hashing. We divide the hashing algorithms two main categories: locality sensitive hashing, which designs hash functions without exploring the data distribution and learning to hash, which learns hash functions according the data distribution, and review them from various aspects, including hash function design and distance measure and search scheme in the hash coding space

    Aligning Language Models with Preferences through f-divergence Minimization

    Full text link
    Aligning language models with preferences can be posed as approximating a target distribution representing some desired behavior. Existing approaches differ both in the functional form of the target distribution and the algorithm used to approximate it. For instance, Reinforcement Learning from Human Feedback (RLHF) corresponds to minimizing a reverse KL from an implicit target distribution arising from a KL penalty in the objective. On the other hand, Generative Distributional Control (GDC) has an explicit target distribution and minimizes a forward KL from it using the Distributional Policy Gradient (DPG) algorithm. In this paper, we propose a new approach, f-DPG, which allows the use of any f-divergence to approximate any target distribution. f-DPG unifies both frameworks (RLHF, GDC) and the approximation methods (DPG, RL with KL penalties). We show the practical benefits of various choices of divergence objectives and demonstrate that there is no universally optimal objective but that different divergences are good for approximating different targets. For instance, we discover that for GDC, the Jensen-Shannon divergence frequently outperforms forward KL divergence by a wide margin, leading to significant improvements over prior work

    Efficient Learning Framework for Training Deep Learning Models with Limited Supervision

    Get PDF
    In recent years, deep learning has shown tremendous success in different applications, however these modes mostly need a large labeled dataset for training their parameters. In this work, we aim to explore the potentials of efficient learning frameworks for training deep models on different problems in the case of limited supervision or noisy labels. For the image clustering problem, we introduce a new deep convolutional autoencoder with an unsupervised learning framework. We employ a relative entropy minimization as the clustering objective regularized by the frequency of cluster assignments and a reconstruction loss. In the case of noisy labels obtained by crowdsourcing platforms, we proposed a novel deep hybrid model for sentiment analysis of text data like tweets based on noisy crowd labels. The proposed model consists of a crowdsourcing aggregation model and a deep text autoencoder. We combine these sub-models based on a probabilistic framework rather than a heuristic way, and derive an efficient optimization algorithm to jointly solve the corresponding problem. In order to improve the performance of unsupervised deep hash functions on image similarity search in big datasets, we adopt generative adversarial networks to propose a new deep image retrieval model, where the adversarial loss is employed as a data-dependent regularization in our objective function. We also introduce a balanced self-paced learning algorithm for training a GAN-based model for image clustering, where the input samples are gradually included into training from easy to difficult, while the diversity of selected samples from all clusters are also considered. In addition, we explore adopting discriminative approaches for unsupervised visual representation learning rather than the generative algorithms, such as maximizing the mutual information between an input image and its representation and a contrastive loss for decreasing the distance between the representations of original and augmented image data

    곡동 λŒ€μ‘°μ  ν•™μŠ΅μ„ μ΄μš©ν•œ 비지도 도메인 적응 기법 연ꡬ

    Get PDF
    ν•™μœ„λ…Όλ¬Έ (석사) -- μ„œμšΈλŒ€ν•™κ΅ λŒ€ν•™μ› : κ³΅κ³ΌλŒ€ν•™ 전기·정보곡학뢀, 2021. 2. μœ€μ„±λ‘œ.Domain adaptation is introduced to exploit the label information of source domain when labels are not available for target domain. Previous methods minimized domain discrepancy in a latent space to enable transfer learning. These studies are based on the theoretical analysis that the target error is upper bounded by the sum of source error, the domain discrepancy, and the joint error of the ideal hypothesis. However, feature discriminability is sacrificed while enhancing the feature transferability by matching marginal distributions. In particular, the ideal joint hypothesis error in the target error upper bound, which was previously considered to be minute, has been found to be significant, impairing its theoretical guarantee. In this paper, to manage the joint error, we propose an alternative upper bound on the target error that explicitly considers it. Based on the theoretical analysis, we suggest a joint optimization framework that combines the source and target domains. To minimize the joint error, we further introduce Joint Contrastive Learning (JCL) that finds class-level discriminative features. With a solid theoretical framework, JCL employs contrastive loss to maximize the mutual information between a feature and its label, which is equivalent to maximizing the Jensen-Shannon divergence between conditional distributions. Extensive experiments on domain adaptation datasets demonstrate that JCL outperforms existing state-of-the-art methods.도메인 적응 기법은 νƒ€κ²Ÿ λ„λ©”μΈμ˜ 라벨 정보가 μ—†λŠ” μƒν™©μ—μ„œ λΉ„μŠ·ν•œ 도메인인 μ†ŒμŠ€ λ„λ©”μΈμ˜ 라벨 정보λ₯Ό ν™œμš©ν•˜κΈ° μœ„ν•΄ κ°œλ°œλ˜μ—ˆλ‹€. 기쑴의 방법둠듀은 잠재 κ³΅κ°„μ—μ„œ 도메인듀 μ‚¬μ΄μ˜ 뢄포 차이λ₯Ό μ€„μž„μœΌλ‘œμ¨ 전이 ν•™μŠ΅μ΄ κ°€λŠ₯ν•˜κ²Œ ν•˜μ˜€λ‹€. μ΄λŸ¬ν•œ 기법듀은 μ†ŒμŠ€ λ„λ©”μΈμ˜ μ—λŸ¬μœ¨, 도메인 κ°„ 뢄포 차이, 그리고 μ–‘ λ„λ©”μΈμ—μ„œ 이상적인 λΆ„λ₯˜κΈ°μ˜ μ—λŸ¬μœ¨μ˜ 합이 νƒ€κ²Ÿ λ„λ©”μΈμ˜ μ—λŸ¬μœ¨μ˜ 상계가 λœλ‹€λŠ” 이둠을 λ°”νƒ•μœΌλ‘œ ν•œλ‹€. κ·ΈλŸ¬λ‚˜ 도메인듀 μ‚¬μ΄μ˜ 뢄포 차이λ₯Ό μ€„μ΄λŠ” 방법듀은 λ™μ‹œμ— 잠재 κ³΅κ°„μ—μ„œ μ„œλ‘œ λ‹€λ₯Έ 라벨을 κ°–λŠ” 데이터듀 μ‚¬μ΄μ˜ ꡬ별성을 κ°μ†Œμ‹œμΌ°λ‹€. 특히, μž‘μ„ 것이라 μƒκ°λ˜λ˜ μ–‘ λ„λ©”μΈμ—μ„œ 이상적인 λΆ„λ₯˜κΈ°μ˜ μ—λŸ¬μœ¨μ΄ 큰 κ²ƒμœΌλ‘œ λ‚˜νƒ€λ‚¬λ‹€. λ³Έ λ…Όλ¬Έμ—μ„œλŠ” 기쑴의 μ΄λ‘ μ—μ„œλŠ” 닀루지 μ•Šμ€ μ–‘ λ„λ©”μΈμ—μ„œ λΆ„λ₯˜κΈ°μ˜ μ—λŸ¬μœ¨μ„ μ‘°μ ˆν•  수 μžˆκ²Œν•˜κΈ° μœ„ν•΄ μƒˆλ‘œμš΄ 이둠을 μ œμ‹œν•œλ‹€. 이 이둠적 배경을 λ°”νƒ•μœΌλ‘œ μ†ŒμŠ€ 도메인과 νƒ€κ²Ÿ 도메인을 ν•¨κ»˜ ν•™μŠ΅ν•˜λŠ” 곡동 λŒ€μ‘°μ  방법을 μ†Œκ°œν•œλ‹€. λ³Έ 곡동 λŒ€μ‘°μ  ν•™μŠ΅ λ°©λ²•μ—μ„œλŠ” 각 λΌλ²¨λ³„λ‘œ κ΅¬λΆ„λ˜λŠ” 잠재 곡간을 ν•™μŠ΅ν•˜κΈ° μœ„ν•΄ 각 λ°μ΄ν„°μ˜ νŠΉμ§•κ³Ό 라벨 μ‚¬μ΄μ˜ μƒν˜Έ μ •λ³΄λŸ‰μ„ μ΅œλŒ€ν™”ν•œλ‹€. 이 각 λ°μ΄ν„°μ˜ νŠΉμ§•κ³Ό 라벨 μ‚¬μ΄μ˜ μƒν˜Έ μ •λ³΄λŸ‰μ€ 각 라벨 뢄포 μ‚¬μ΄μ˜ μ  μ„Ό-샀논 거리와 κ°™μœΌλ―€λ‘œ 이λ₯Ό μ΅œλŒ€ν™”ν•˜λŠ” 것은 곧 라벨듀이 잘 κ΅¬λ³„λ˜λŠ” 잠재 곡간을 ν•™μŠ΅ν•˜λŠ” 것이닀. λ§ˆμ§€λ§‰μœΌλ‘œ 곡동 λŒ€μ‘°μ  ν•™μŠ΅ 방법을 μ—¬λŸ¬ 데이터 셋에 μ μš©ν•˜μ—¬ κΈ°μ‘΄ 방법둠듀과 λΉ„κ΅ν•˜μ˜€λ‹€.1 Introduction 1 2 Background 4 2.1 Domain Adaptation . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 2.1.1 Problem Setting and Notations . . . . . . . . . . . . . . . . . 4 2.1.2 Theoretical Background . . . . . . . . . . . . . . . . . . . . 5 2.2 Approaches for Domain Adaptation . . . . . . . . . . . . . . . . . . 6 2.2.1 Marginal Distribution Alignment Based Approaches . . . . . 6 2.2.2 Conditional Distribution Matching Approaches . . . . . . . . 7 2.3 Contrastive Learning . . . . . . . . . . . . . . . . . . . . . . . . . . 8 3 Method 10 3.1 An Alternative Upper Bound . . . . . . . . . . . . . . . . . . . . . . 10 3.2 Joint Contrastive Learning . . . . . . . . . . . . . . . . . . . . . . . 14 3.2.1 Theoretical Guarantees . . . . . . . . . . . . . . . . . . . . . 14 3.2.2 Generalization to Multiclass Classification . . . . . . . . . . 17 3.2.3 Training Procedure . . . . . . . . . . . . . . . . . . . . . . . 19 4 Experiments 24 4.1 Datasets and Baselines . . . . . . . . . . . . . . . . . . . . . . . . . 24 4.2 Implementation Details . . . . . . . . . . . . . . . . . . . . . . . . . 26 4.3 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 4.4 Ablation Studies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 5 Conclusion 35 Abstract (In Korean) 45Maste
    • …
    corecore