3,572 research outputs found

    Tunable Quantum Beam Splitters for Coherent Manipulation of a Solid-State Tripartite Qubit System

    Get PDF
    Coherent control of quantum states is at the heart of implementing solid-state quantum processors and testing quantum mechanics at the macroscopic level. Despite significant progress made in recent years in controlling single- and bi-partite quantum systems, coherent control of quantum wave function in multipartite systems involving artificial solid-state qubits has been hampered due to the relatively short decoherence time and lacking of precise control methods. Here we report the creation and coherent manipulation of quantum states in a tripartite quantum system, which is formed by a superconducting qubit coupled to two microscopic two-level systems (TLSs). The avoided crossings in the system's energy-level spectrum due to the qubit-TLS interaction act as tunable quantum beam splitters of wave functions. Our result shows that the Landau-Zener-St\"{u}ckelberg interference has great potential in the precise control of the quantum states in the tripartite system.Comment: 24 pages, 3 figure

    On the Over-Memorization During Natural, Robust and Catastrophic Overfitting

    Full text link
    Overfitting negatively impacts the generalization ability of deep neural networks (DNNs) in both natural and adversarial training. Existing methods struggle to consistently address different types of overfitting, typically designing strategies that focus separately on either natural or adversarial patterns. In this work, we adopt a unified perspective by solely focusing on natural patterns to explore different types of overfitting. Specifically, we examine the memorization effect in DNNs and reveal a shared behaviour termed over-memorization, which impairs their generalization capacity. This behaviour manifests as DNNs suddenly becoming high-confidence in predicting certain training patterns and retaining a persistent memory for them. Furthermore, when DNNs over-memorize an adversarial pattern, they tend to simultaneously exhibit high-confidence prediction for the corresponding natural pattern. These findings motivate us to holistically mitigate different types of overfitting by hindering the DNNs from over-memorization natural patterns. To this end, we propose a general framework, Distraction Over-Memorization (DOM), which explicitly prevents over-memorization by either removing or augmenting the high-confidence natural patterns. Extensive experiments demonstrate the effectiveness of our proposed method in mitigating overfitting across various training paradigms

    Improved Noisy Student Training for Automatic Speech Recognition

    Full text link
    Recently, a semi-supervised learning method known as "noisy student training" has been shown to improve image classification performance of deep networks significantly. Noisy student training is an iterative self-training method that leverages augmentation to improve network performance. In this work, we adapt and improve noisy student training for automatic speech recognition, employing (adaptive) SpecAugment as the augmentation method. We find effective methods to filter, balance and augment the data generated in between self-training iterations. By doing so, we are able to obtain word error rates (WERs) 4.2%/8.6% on the clean/noisy LibriSpeech test sets by only using the clean 100h subset of LibriSpeech as the supervised set and the rest (860h) as the unlabeled set. Furthermore, we are able to achieve WERs 1.7%/3.4% on the clean/noisy LibriSpeech test sets by using the unlab-60k subset of LibriLight as the unlabeled set for LibriSpeech 960h. We are thus able to improve upon the previous state-of-the-art clean/noisy test WERs achieved on LibriSpeech 100h (4.74%/12.20%) and LibriSpeech (1.9%/4.1%).Comment: 5 pages, 5 figures, 4 tables; v2: minor revisions, reference adde
    • …
    corecore