4,677 research outputs found

    Hierarchically Clustered Representation Learning

    Full text link
    The joint optimization of representation learning and clustering in the embedding space has experienced a breakthrough in recent years. In spite of the advance, clustering with representation learning has been limited to flat-level categories, which often involves cohesive clustering with a focus on instance relations. To overcome the limitations of flat clustering, we introduce hierarchically-clustered representation learning (HCRL), which simultaneously optimizes representation learning and hierarchical clustering in the embedding space. Compared with a few prior works, HCRL firstly attempts to consider a generation of deep embeddings from every component of the hierarchy, not just leaf components. In addition to obtaining hierarchically clustered embeddings, we can reconstruct data by the various abstraction levels, infer the intrinsic hierarchical structure, and learn the level-proportion features. We conducted evaluations with image and text domains, and our quantitative analyses showed competent likelihoods and the best accuracies compared with the baselines.Comment: 10 pages, 7 figures, Under review as a conference pape

    LFS-GAN: Lifelong Few-Shot Image Generation

    Full text link
    We address a challenging lifelong few-shot image generation task for the first time. In this situation, a generative model learns a sequence of tasks using only a few samples per task. Consequently, the learned model encounters both catastrophic forgetting and overfitting problems at a time. Existing studies on lifelong GANs have proposed modulation-based methods to prevent catastrophic forgetting. However, they require considerable additional parameters and cannot generate high-fidelity and diverse images from limited data. On the other hand, the existing few-shot GANs suffer from severe catastrophic forgetting when learning multiple tasks. To alleviate these issues, we propose a framework called Lifelong Few-Shot GAN (LFS-GAN) that can generate high-quality and diverse images in lifelong few-shot image generation task. Our proposed framework learns each task using an efficient task-specific modulator - Learnable Factorized Tensor (LeFT). LeFT is rank-constrained and has a rich representation ability due to its unique reconstruction technique. Furthermore, we propose a novel mode seeking loss to improve the diversity of our model in low-data circumstances. Extensive experiments demonstrate that the proposed LFS-GAN can generate high-fidelity and diverse images without any forgetting and mode collapse in various domains, achieving state-of-the-art in lifelong few-shot image generation task. Surprisingly, we find that our LFS-GAN even outperforms the existing few-shot GANs in the few-shot image generation task. The code is available at Github.Comment: 20 pages, 19 figures, 14 tables, ICCV 2023 Poste

    Single-Copy Certification of Two-Qubit Gates without Entanglement

    Full text link
    A quantum state transformation can be generally approximated by single- and two-qubit gates. This, however, does not hold with noisy intermediate-scale quantum technologies due to the errors appearing in the gate operations, where errors of two-qubit gates such as controlled-NOT and SWAP operations are dominated. In this work, we present a cost efficient single-copy certification for a realization of a two-qubit gate in the presence of depolarization noise, where it is aimed to identify if the realization is noise-free, or not. It is shown that entangled resources such as entangled states and a joint measurement are not necessary for the purpose, i.e., a noise-free two-qubit gate is not needed to certify an implementation of a two-qubit gate. A proof-of-principle demonstration is presented with photonic qubits.Comment: 8 pages. arXiv admin note: text overlap with arXiv:1812.0208

    Vitamin D deficiency in infants aged 1 to 6 months

    Get PDF

    Virtual Reality In Enhancing Internet-Based Education

    Get PDF
    Until recently, Internet-based education is fundamentally two-dimensional (2-D). This difference from the three-dimensional (3-D) world that human beings experience and learn provokes that 3-D learning environments should be included into the Internet-based education. It is envisaged that Virtual Reality (VR) has the capability to provide for the 3-D learning environments. The objectives of this study are to investigate the capability of VR in enhancing Internet-based education and to investigate the process of VR model or Virtual Environment (VE) development. As Virtual Reality Modeling language (VRML) is the non-proprietary 3 -D format to represent VR on the Internet and its current version (2.0) enables dynamic, interactive models to be developed thes

    Perception-Oriented Single Image Super-Resolution using Optimal Objective Estimation

    Full text link
    Single-image super-resolution (SISR) networks trained with perceptual and adversarial losses provide high-contrast outputs compared to those of networks trained with distortion-oriented losses, such as L1 or L2. However, it has been shown that using a single perceptual loss is insufficient for accurately restoring locally varying diverse shapes in images, often generating undesirable artifacts or unnatural details. For this reason, combinations of various losses, such as perceptual, adversarial, and distortion losses, have been attempted, yet it remains challenging to find optimal combinations. Hence, in this paper, we propose a new SISR framework that applies optimal objectives for each region to generate plausible results in overall areas of high-resolution outputs. Specifically, the framework comprises two models: a predictive model that infers an optimal objective map for a given low-resolution (LR) input and a generative model that applies a target objective map to produce the corresponding SR output. The generative model is trained over our proposed objective trajectory representing a set of essential objectives, which enables the single network to learn various SR results corresponding to combined losses on the trajectory. The predictive model is trained using pairs of LR images and corresponding optimal objective maps searched from the objective trajectory. Experimental results on five benchmarks show that the proposed method outperforms state-of-the-art perception-driven SR methods in LPIPS, DISTS, PSNR, and SSIM metrics. The visual results also demonstrate the superiority of our method in perception-oriented reconstruction. The code and models are available at https://github.com/seungho-snu/SROOE.Comment: Code and trained models will be available at https://github.com/seungho-snu/SROO
    corecore