96 research outputs found

    DC-Informative Joint Color-Frequency Modulation for Visible Light Communications

    Full text link
    In this paper, we consider the problem of constellation design for a visible light communication (VLC) system using red/green/blue light-emitting diodes (RGB LED), and propose a method termed DC-informative joint color-frequency modulation (DCI-JCFM). This method jointly utilizes available diversity resources including different optical wavelengths, multiple baseband subcarriers, and adaptive DC-bias. Constellation is designed in a high dimensional space, where the compact sphere packing advantage over lower dimensional counterparts is utilized. Taking into account multiple practical illumination constraints, a non-convex optimization problem is formulated, seeking the least error rate with a fixed spectral efficiency. The proposed scheme is compared with a decoupled scheme, where constellation is designed separately for each LED. Notable gains for DCI-JCFM are observed through simulations where balanced, unbalanced and very unbalanced color illuminations are considered.Comment: submitted to Journal of Lightwave Technology, Aug. 5th 201

    Weak Alignment Supervision from Hybrid Model Improves End-to-end ASR

    Full text link
    In this paper, we aim to create weak alignment supervision from an existing hybrid system to aid the end-to-end modeling of automatic speech recognition. Towards this end, we use the existing hybrid ASR system to produce triphone alignments of the training audios. We then create a cross-entropy loss at a certain layer of the encoder using the derived alignments. In contrast to the general one-hot cross-entropy losses, here we use a cross-entropy loss with a label smoothing parameter to regularize the supervision. As a comparison, we also conduct the experiments with one-hot cross-entropy losses and CTC losses with loss weighting. The results show that placing the weak alignment supervision with the label smoothing parameter of 0.5 at the third encoder layer outperforms the other two approaches and leads to about 5\% relative WER reduction on the TED-LIUM 2 dataset over the baseline. We see similar improvements when applying the method out-of-the-box on a Tagalog end-to-end ASR system.Comment: 7 pages, 7 figures, and 5 table

    Improving Language Model Integration for Neural Machine Translation

    Full text link
    The integration of language models for neural machine translation has been extensively studied in the past. It has been shown that an external language model, trained on additional target-side monolingual data, can help improve translation quality. However, there has always been the assumption that the translation model also learns an implicit target-side language model during training, which interferes with the external language model at decoding time. Recently, some works on automatic speech recognition have demonstrated that, if the implicit language model is neutralized in decoding, further improvements can be gained when integrating an external language model. In this work, we transfer this concept to the task of machine translation and compare with the most prominent way of including additional monolingual data - namely back-translation. We find that accounting for the implicit language model significantly boosts the performance of language model fusion, although this approach is still outperformed by back-translation.Comment: accepted at ACL2023 (Findings

    Adaptive estimation and control of MR damper for semi-active suspension systems

    Get PDF

    Self-Normalized Importance Sampling for Neural Language Modeling

    Full text link
    To mitigate the problem of having to traverse over the full vocabulary in the softmax normalization of a neural language model, sampling-based training criteria are proposed and investigated in the context of large vocabulary word-based neural language models. These training criteria typically enjoy the benefit of faster training and testing, at a cost of slightly degraded performance in terms of perplexity and almost no visible drop in word error rate. While noise contrastive estimation is one of the most popular choices, recently we show that other sampling-based criteria can also perform well, as long as an extra correction step is done, where the intended class posterior probability is recovered from the raw model outputs. In this work, we propose self-normalized importance sampling. Compared to our previous work, the criteria considered in this work are self-normalized and there is no need to further conduct a correction step. Through self-normalized language model training as well as lattice rescoring experiments, we show that our proposed self-normalized importance sampling is competitive in both research-oriented and production-oriented automatic speech recognition tasks.Comment: Accepted at INTERSPEECH 202
    corecore