13 research outputs found

    Contextual Linear Bandits under Noisy Features: Towards Bayesian Oracles

    Full text link
    We study contextual linear bandit problems under uncertainty on features; they are noisy with missing entries. To address the challenges from the noise, we analyze Bayesian oracles given observed noisy features. Our Bayesian analysis finds that the optimal hypothesis can be far from the underlying realizability function, depending on noise characteristics, which is highly non-intuitive and does not occur for classical noiseless setups. This implies that classical approaches cannot guarantee a non-trivial regret bound. We thus propose an algorithm aiming at the Bayesian oracle from observed information under this model, achieving O~(dT)\tilde{O}(d\sqrt{T}) regret bound with respect to feature dimension dd and time horizon TT. We demonstrate the proposed algorithm using synthetic and real-world datasets.Comment: 30 page

    Efficient Parallel Audio Generation using Group Masked Language Modeling

    Full text link
    We present a fast and high-quality codec language model for parallel audio generation. While SoundStorm, a state-of-the-art parallel audio generation model, accelerates inference speed compared to autoregressive models, it still suffers from slow inference due to iterative sampling. To resolve this problem, we propose Group-Masked Language Modeling~(G-MLM) and Group Iterative Parallel Decoding~(G-IPD) for efficient parallel audio generation. Both the training and sampling schemes enable the model to synthesize high-quality audio with a small number of iterations by effectively modeling the group-wise conditional dependencies. In addition, our model employs a cross-attention-based architecture to capture the speaker style of the prompt voice and improves computational efficiency. Experimental results demonstrate that our proposed model outperforms the baselines in prompt-based audio generation.Comment: This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessibl

    FedSoL: Bridging Global Alignment and Local Generality in Federated Learning

    Full text link
    Federated Learning (FL) aggregates locally trained models from individual clients to construct a global model. While FL enables learning a model with data privacy, it often suffers from significant performance degradation when client data distributions are heterogeneous. Many previous FL algorithms have addressed this issue by introducing various proximal restrictions. These restrictions aim to encourage global alignment by constraining the deviation of local learning from the global objective. However, they inherently limit local learning by interfering with the original local objectives. Recently, an alternative approach has emerged to improve local learning generality. By obtaining local models within a smooth loss landscape, this approach mitigates conflicts among different local objectives of the clients. Yet, it does not ensure stable global alignment, as local learning does not take the global objective into account. In this study, we propose Federated Stability on Learning (FedSoL), which combines both the concepts of global alignment and local generality. In FedSoL, the local learning seeks a parameter region robust against proximal perturbations. This strategy introduces an implicit proximal restriction effect in local learning while maintaining the original local objective for parameter update. Our experiments show that FedSoL consistently achieves state-of-the-art performance on various setups

    FedDr+: Stabilizing Dot-regression with Global Feature Distillation for Federated Learning

    Full text link
    Federated Learning (FL) has emerged as a pivotal framework for the development of effective global models (global FL) or personalized models (personalized FL) across clients with heterogeneous, non-iid data distribution. A key challenge in FL is client drift, where data heterogeneity impedes the aggregation of scattered knowledge. Recent studies have tackled the client drift issue by identifying significant divergence in the last classifier layer. To mitigate this divergence, strategies such as freezing the classifier weights and aligning the feature extractor accordingly have proven effective. Although the local alignment between classifier and feature extractor has been studied as a crucial factor in FL, we observe that it may lead the model to overemphasize the observed classes within each client. Thus, our objectives are twofold: (1) enhancing local alignment while (2) preserving the representation of unseen class samples. This approach aims to effectively integrate knowledge from individual clients, thereby improving performance for both global and personalized FL. To achieve this, we introduce a novel algorithm named FedDr+, which empowers local model alignment using dot-regression loss. FedDr+ freezes the classifier as a simplex ETF to align the features and improves aggregated global models by employing a feature distillation mechanism to retain information about unseen/missing classes. Consequently, we provide empirical evidence demonstrating that our algorithm surpasses existing methods that use a frozen classifier to boost alignment across the diverse distribution

    High methane combustion activity of PdO/CeO2–ZrO2–NiO/γ-Al2O3 catalysts

    No full text
    PdO/CeO2–ZrO2–NiO/γ-Al2O3 catalysts were prepared for the combustion of methane at moderate temperatures. The introduction of a small amount of NiO within the cubic fluorite CeO2–ZrO2 structure as a promoter effectively enhanced the oxygen release and storage abilities of the catalysts, thereby achieving the complete oxidation of methane. The catalyst with the highest activity for methane combustion was 11.3 mass% PdO/20 mass% Ce0.64Zr0.16Ni0.2O1.9/γ-Al2O3, and efficient combustion was realized at a temperature as low as 300 °C

    Preservation of Global Knowledge by Not-True Distillation in Federated Learning

    Full text link
    In federated learning, a strong global model is collaboratively learned by aggregating clients' locally trained models. Although this precludes the need to access clients' data directly, the global model's convergence often suffers from data heterogeneity. This study starts from an analogy to continual learning and suggests that forgetting could be the bottleneck of federated learning. We observe that the global model forgets the knowledge from previous rounds, and the local training induces forgetting the knowledge outside of the local distribution. Based on our findings, we hypothesize that tackling down forgetting will relieve the data heterogeneity problem. To this end, we propose a novel and effective algorithm, Federated Not-True Distillation (FedNTD), which preserves the global perspective on locally available data only for the not-true classes. In the experiments, FedNTD shows state-of-the-art performance on various setups without compromising data privacy or incurring additional communication costs.Comment: Under revie

    Three-dimensional surface lattice plasmon resonance effect from plasmonic inclined nanostructures via one-step stencil lithography

    No full text
    Plasmonic nanostructures allow the manipulation and confinement of optical fields on the sub-wavelength scale. The local field enhancement and environmentally sensitive resonance characteristics provided by these nanostructures are of high importance for biological and chemical sensing. Recently, surface lattice plasmon resonance (SLR) research has attracted much interest because of its superior quality factor (Q-factor) compared to that of localized surface plasmon resonances (LSPR), which is facilitated by resonant plasmonic mode coupling between individual nanostructures over a large area. This advantage can be further enhanced by utilizing asymmetric 3D structures rather than low-height (typically height < ∼60 nm) structure arrays, which results in stronger coupling due to an increased mode volume. However, fabricating 3D, high-aspect ratio, symmetry-breaking structures is a complex and challenging process even with state-of-the-art fabrication technology. Here, we report a plasmonic metasurface of 3D inclined structures produced via commercial TEM grid–based stencil lithography with a Q-factor of 101.6, a refractive index sensitivity of 291 nm/RIU, and a figure of merit (FOM) of 44.7 in the visible wavelength range at a refractive index of 1.5 by utilizing the 3D SLR enhancement effect, which exceeds the performance of most LSPR systems (Q < ∼10). The symmetry-breaking 3D inclined structures that are fabricated by electron beam evaporation at an angle increase the polarizability of the metasurface and the directionality of the diffractively scattered radiative field responsible for SLR mode coupling. Additionally, we explore the role of spatial coherence in facilitating the SLR effect and thus a high-Q plasmonic response from the nanostructures. Our work demonstrates the feasibility of producing 3D inclined structure arrays with pronounced SLR enhancement for high biological sensitivity by utilizing the previously unexplored inclined stencil lithography, which opens the way to fabricate highly sensitive plasmonic metasurfaces with this novel simple technique
    corecore