768 research outputs found
Lower bounds on the error probability of multiple quantum channel discrimination by the Bures angle and the trace distance
Quantum channel discrimination is a fundamental problem in quantum
information science. In this study, we consider general quantum channel
discrimination problems, and derive the lower bounds of the error probability.
Our lower bounds are based on the triangle inequalities of the Bures angle and
the trace distance. As a consequence of the lower bound based on the Bures
angle, we prove the optimality of Grover's search if the number of marked
elements is fixed to some integer . This result generalizes Zalka's result
for . We also present several numerical results in which our lower bounds
based on the trace distance outperform recently obtained lower bounds.Comment: 14 pages, 6 figure
ヒト脂肪組織由来再生細胞を用いたラット骨格筋損傷モデルの骨格筋再生の促進
内容の要旨 , 審査の要旨広島大学(Hiroshima University)博士(医学)Philosophy in Medical Sciencedoctora
Recommended from our members
Controlling a Van Hove singularity and Fermi surface topology at a complex oxide heterostructure interface.
The emergence of saddle-point Van Hove singularities (VHSs) in the density of states, accompanied by a change in Fermi surface topology, Lifshitz transition, constitutes an ideal ground for the emergence of different electronic phenomena, such as superconductivity, pseudo-gap, magnetism, and density waves. However, in most materials the Fermi level, [Formula: see text], is too far from the VHS where the change of electronic topology takes place, making it difficult to reach with standard chemical doping or gating techniques. Here, we demonstrate that this scenario can be realized at the interface between a Mott insulator and a band insulator as a result of quantum confinement and correlation enhancement, and easily tuned by fine control of layer thickness and orbital occupancy. These results provide a tunable pathway for Fermi surface topology and VHS engineering of electronic phases
Personalized Federated Learning with Multi-branch Architecture
Federated learning (FL) is a decentralized machine learning technique that
enables multiple clients to collaboratively train models without requiring
clients to reveal their raw data to each other. Although traditional FL trains
a single global model with average performance among clients, statistical data
heterogeneity across clients has resulted in the development of personalized FL
(PFL), which trains personalized models with good performance on each client's
data. A key challenge with PFL is how to facilitate clients with similar data
to collaborate more in a situation where each client has data from complex
distribution and cannot determine one another's distribution. In this paper, we
propose a new PFL method (pFedMB) using multi-branch architecture, which
achieves personalization by splitting each layer of a neural network into
multiple branches and assigning client-specific weights to each branch. We also
design an aggregation method to improve the communication efficiency and the
model performance, with which each branch is globally updated with weighted
averaging by client-specific weights assigned to the branch. pFedMB is simple
but effective in facilitating each client to share knowledge with similar
clients by adjusting the weights assigned to each branch. We experimentally
show that pFedMB performs better than the state-of-the-art PFL methods using
the CIFAR10 and CIFAR100 datasets.Comment: Accepted by IJCNN 202
Heterogeneous Domain Adaptation with Positive and Unlabeled Data
Heterogeneous unsupervised domain adaptation (HUDA) is the most challenging
domain adaptation setting where the feature spaces of source and target domains
are heterogeneous, and the target domain has only unlabeled data. Existing HUDA
methods assume that both positive and negative examples are available in the
source domain, which may not be satisfied in some real applications. This paper
addresses a new challenging setting called positive and unlabeled heterogeneous
unsupervised domain adaptation (PU-HUDA), a HUDA setting where the source
domain only has positives. PU-HUDA can also be viewed as an extension of PU
learning where the positive and unlabeled examples are sampled from different
domains. A naive combination of existing HUDA and PU learning methods is
ineffective in PU-HUDA due to the gap in label distribution between the source
and target domains. To overcome this issue, we propose a novel method,
predictive adversarial domain adaptation (PADA), which can predict likely
positive examples from the unlabeled target data and simultaneously align the
feature spaces to reduce the distribution divergence between the whole source
data and the likely positive target data. PADA achieves this by a unified
adversarial training framework for learning a classifier to predict positive
examples and a feature transformer to transform the target feature space to
that of the source. Specifically, they are both trained to fool a common
discriminator that determines whether the likely positive examples are from the
target or source domain. We experimentally show that PADA outperforms several
baseline methods, such as the naive combination of HUDA and PU learning.Comment: Accepted by IEEE Big Data 2023 as a regular pape
Validating Variational Bayes Linear Regression Method With Multi-Central Datasets.
PurposeTo validate the prediction accuracy of variational Bayes linear regression (VBLR) with two datasets external to the training dataset.MethodThe training dataset consisted of 7268 eyes of 4278 subjects from the University of Tokyo Hospital. The Japanese Archive of Multicentral Databases in Glaucoma (JAMDIG) dataset consisted of 271 eyes of 177 patients, and the Diagnostic Innovations in Glaucoma Study (DIGS) dataset includes 248 eyes of 173 patients, which were used for validation. Prediction accuracy was compared between the VBLR and ordinary least squared linear regression (OLSLR). First, OLSLR and VBLR were carried out using total deviation (TD) values at each of the 52 test points from the second to fourth visual fields (VFs) (VF2-4) to 2nd to 10th VF (VF2-10) of each patient in JAMDIG and DIGS datasets, and the TD values of the 11th VF test were predicted every time. The predictive accuracy of each method was compared through the root mean squared error (RMSE) statistic.ResultsOLSLR RMSEs with the JAMDIG and DIGS datasets were between 31 and 4.3 dB, and between 19.5 and 3.9 dB. On the other hand, VBLR RMSEs with JAMDIG and DIGS datasets were between 5.0 and 3.7, and between 4.6 and 3.6 dB. There was statistically significant difference between VBLR and OLSLR for both datasets at every series (VF2-4 to VF2-10) (P < 0.01 for all tests). However, there was no statistically significant difference in VBLR RMSEs between JAMDIG and DIGS datasets at any series of VFs (VF2-2 to VF2-10) (P > 0.05).ConclusionsVBLR outperformed OLSLR to predict future VF progression, and the VBLR has a potential to be a helpful tool at clinical settings
- …