2,149 research outputs found

    Laser cooling of trapped ytterbium ions with an ultraviolet diode laser

    Full text link
    We demonstrate an ultraviolet diode laser system for cooling of trapped ytterbium ions. The laser power and linewidth are comparable to previous systems based on resonant frequency doubling, but the system is simpler, more robust, and less expensive. We use the laser system to cool small numbers of ytterbium ions confined in a linear Paul trap. From the observed spectra, we deduce final temperatures < 270 mK.Comment: submitted to Opt. Let

    Mitochondria and G-quadruplex evolution: an intertwined relationship

    Get PDF
    G-quadruplexes (G4s) are non-canonical structures formed in guanine (G)-rich sequences through stacked G tetrads by Hoogsteen hydrogen bonding. Several studies have demonstrated the existence of G4s in the genome of various organisms, including humans, and have proposed that G4s have a regulatory role in various cellular functions. However, little is known regarding the dissemination of G4s in mitochondria. In this review, we report the observation that the number of potential G4-forming sequences in the mitochondrial genome increases with the evolutionary complexity of different species, suggesting that G4s have a beneficial role in higher-order organisms. We also discuss the possible function of G4s in mitochondrial (mt)DNA and long noncoding (lnc)RNA and their role in various biological processes

    青斑核-大脳皮質系電気活動に対するノルアドレナリン性修飾

    Get PDF
    取得学位 : 博士(医学), 学位授与番号 : 医博甲第912号, 学位授与年月日:平成1年6月30日,学位授与年:198

    How does Disagreement Help Generalization against Label Corruption?

    Full text link
    Learning with noisy labels is one of the hottest problems in weakly-supervised learning. Based on memorization effects of deep neural networks, training on small-loss instances becomes very promising for handling noisy labels. This fosters the state-of-the-art approach "Co-teaching" that cross-trains two deep neural networks using the small-loss trick. However, with the increase of epochs, two networks converge to a consensus and Co-teaching reduces to the self-training MentorNet. To tackle this issue, we propose a robust learning paradigm called Co-teaching+, which bridges the "Update by Disagreement" strategy with the original Co-teaching. First, two networks feed forward and predict all data, but keep prediction disagreement data only. Then, among such disagreement data, each network selects its small-loss data, but back propagates the small-loss data from its peer network and updates its own parameters. Empirical results on benchmark datasets demonstrate that Co-teaching+ is much superior to many state-of-the-art methods in the robustness of trained models

    Towards Mixture Proportion Estimation without Irreducibility

    Full text link
    \textit{Mixture proportion estimation} (MPE) is a fundamental problem of practical significance, where we are given data from only a \textit{mixture} and one of its two \textit{components} to identify the proportion of each component. All existing MPE methods that are distribution-independent explicitly or implicitly rely on the \textit{irreducible} assumption---the unobserved component is not a mixture containing the observable component. If this is not satisfied, those methods will lead to a critical estimation bias. In this paper, we propose \textit{Regrouping-MPE} that works without irreducible assumption: it builds a new irreducible MPE problem and solves the new problem. It is worthwhile to change the problem: we prove that if the assumption holds, our method will not affect anything; if the assumption does not hold, the bias from problem changing is less than the bias from violation of the irreducible assumption in the original problem. Experiments show that our method outperforms all state-of-the-art MPE methods on various real-world datasets
    corecore