4,677 research outputs found
Hierarchically Clustered Representation Learning
The joint optimization of representation learning and clustering in the
embedding space has experienced a breakthrough in recent years. In spite of the
advance, clustering with representation learning has been limited to flat-level
categories, which often involves cohesive clustering with a focus on instance
relations. To overcome the limitations of flat clustering, we introduce
hierarchically-clustered representation learning (HCRL), which simultaneously
optimizes representation learning and hierarchical clustering in the embedding
space. Compared with a few prior works, HCRL firstly attempts to consider a
generation of deep embeddings from every component of the hierarchy, not just
leaf components. In addition to obtaining hierarchically clustered embeddings,
we can reconstruct data by the various abstraction levels, infer the intrinsic
hierarchical structure, and learn the level-proportion features. We conducted
evaluations with image and text domains, and our quantitative analyses showed
competent likelihoods and the best accuracies compared with the baselines.Comment: 10 pages, 7 figures, Under review as a conference pape
LFS-GAN: Lifelong Few-Shot Image Generation
We address a challenging lifelong few-shot image generation task for the
first time. In this situation, a generative model learns a sequence of tasks
using only a few samples per task. Consequently, the learned model encounters
both catastrophic forgetting and overfitting problems at a time. Existing
studies on lifelong GANs have proposed modulation-based methods to prevent
catastrophic forgetting. However, they require considerable additional
parameters and cannot generate high-fidelity and diverse images from limited
data. On the other hand, the existing few-shot GANs suffer from severe
catastrophic forgetting when learning multiple tasks. To alleviate these
issues, we propose a framework called Lifelong Few-Shot GAN (LFS-GAN) that can
generate high-quality and diverse images in lifelong few-shot image generation
task. Our proposed framework learns each task using an efficient task-specific
modulator - Learnable Factorized Tensor (LeFT). LeFT is rank-constrained and
has a rich representation ability due to its unique reconstruction technique.
Furthermore, we propose a novel mode seeking loss to improve the diversity of
our model in low-data circumstances. Extensive experiments demonstrate that the
proposed LFS-GAN can generate high-fidelity and diverse images without any
forgetting and mode collapse in various domains, achieving state-of-the-art in
lifelong few-shot image generation task. Surprisingly, we find that our LFS-GAN
even outperforms the existing few-shot GANs in the few-shot image generation
task. The code is available at Github.Comment: 20 pages, 19 figures, 14 tables, ICCV 2023 Poste
Single-Copy Certification of Two-Qubit Gates without Entanglement
A quantum state transformation can be generally approximated by single- and
two-qubit gates. This, however, does not hold with noisy intermediate-scale
quantum technologies due to the errors appearing in the gate operations, where
errors of two-qubit gates such as controlled-NOT and SWAP operations are
dominated. In this work, we present a cost efficient single-copy certification
for a realization of a two-qubit gate in the presence of depolarization noise,
where it is aimed to identify if the realization is noise-free, or not. It is
shown that entangled resources such as entangled states and a joint measurement
are not necessary for the purpose, i.e., a noise-free two-qubit gate is not
needed to certify an implementation of a two-qubit gate. A proof-of-principle
demonstration is presented with photonic qubits.Comment: 8 pages. arXiv admin note: text overlap with arXiv:1812.0208
Virtual Reality In Enhancing Internet-Based Education
Until recently, Internet-based education is fundamentally two-dimensional
(2-D). This difference from the three-dimensional (3-D) world that human beings
experience and learn provokes that 3-D learning environments should be included
into the Internet-based education. It is envisaged that Virtual Reality (VR) has the
capability to provide for the 3-D learning environments.
The objectives of this study are to investigate the capability of VR in
enhancing Internet-based education and to investigate the process of VR model or
Virtual Environment (VE) development. As Virtual Reality Modeling language
(VRML) is the non-proprietary 3 -D format to represent VR on the Internet and its
current version (2.0) enables dynamic, interactive models to be developed thes
Perception-Oriented Single Image Super-Resolution using Optimal Objective Estimation
Single-image super-resolution (SISR) networks trained with perceptual and
adversarial losses provide high-contrast outputs compared to those of networks
trained with distortion-oriented losses, such as L1 or L2. However, it has been
shown that using a single perceptual loss is insufficient for accurately
restoring locally varying diverse shapes in images, often generating
undesirable artifacts or unnatural details. For this reason, combinations of
various losses, such as perceptual, adversarial, and distortion losses, have
been attempted, yet it remains challenging to find optimal combinations. Hence,
in this paper, we propose a new SISR framework that applies optimal objectives
for each region to generate plausible results in overall areas of
high-resolution outputs. Specifically, the framework comprises two models: a
predictive model that infers an optimal objective map for a given
low-resolution (LR) input and a generative model that applies a target
objective map to produce the corresponding SR output. The generative model is
trained over our proposed objective trajectory representing a set of essential
objectives, which enables the single network to learn various SR results
corresponding to combined losses on the trajectory. The predictive model is
trained using pairs of LR images and corresponding optimal objective maps
searched from the objective trajectory. Experimental results on five benchmarks
show that the proposed method outperforms state-of-the-art perception-driven SR
methods in LPIPS, DISTS, PSNR, and SSIM metrics. The visual results also
demonstrate the superiority of our method in perception-oriented
reconstruction. The code and models are available at
https://github.com/seungho-snu/SROOE.Comment: Code and trained models will be available at
https://github.com/seungho-snu/SROO
- …