1,078 research outputs found

    Off-target at-scale Scale Down Model verification of a marketed biopharmaceutical

    Get PDF
    According to authority guidelines, process understanding is considered the key factor for a successful control strategy in submissions for biopharmaceutical production processes. This knowledge is usually gained in scale down model (SDM) experiments and therefore is subject to the ability of the SDM to reliably predict behavior of cell culture processes in manufacturing scale. In order to prove predictability, SDM runs at target conditions are typically compared to manufacturing scale target runs only. At the same time, predictability is also assumed within the whole design space described or even beyond as evaluated in process design studies. The proof at off-target conditions, however, is rarely provided. In this presentation, the author will exhibit an example where a well-established scale down model was used to determine off-target conditions to manufacture clinical material of a marketed product in 10,000 L scale with a deliberately deviant from target but nevertheless well-defined product quality profile. The desired quality attribute profile was achieved in two manufacturing runs with remarkable accuracy providing additional evidence of the SDM predictability. Hence, confidence is increased in the capability of the resultant control system which has been submitted to authorities for a post licensure manufacturing change

    Don't Forget to Fight! : Singapore's History Education and War Commemoration, 1945-2005

    Get PDF
    Ph.DDOCTOR OF PHILOSOPH

    ESPnet-ONNX: Bridging a Gap Between Research and Production

    Full text link
    In the field of deep learning, researchers often focus on inventing novel neural network models and improving benchmarks. In contrast, application developers are interested in making models suitable for actual products, which involves optimizing a model for faster inference and adapting a model to various platforms (e.g., C++ and Python). In this work, to fill the gap between the two, we establish an effective procedure for optimizing a PyTorch-based research-oriented model for deployment, taking ESPnet, a widely used toolkit for speech processing, as an instance. We introduce different techniques to ESPnet, including converting a model into an ONNX format, fusing nodes in a graph, and quantizing parameters, which lead to approximately 1.3-2×\times speedup in various tasks (i.e., ASR, TTS, speech translation, and spoken language understanding) while keeping its performance without any additional training. Our ESPnet-ONNX will be publicly available at https://github.com/espnet/espnet_onnxComment: Accepted to APSIPA ASC 202

    InterMPL: Momentum Pseudo-Labeling with Intermediate CTC Loss

    Full text link
    This paper presents InterMPL, a semi-supervised learning method of end-to-end automatic speech recognition (ASR) that performs pseudo-labeling (PL) with intermediate supervision. Momentum PL (MPL) trains a connectionist temporal classification (CTC)-based model on unlabeled data by continuously generating pseudo-labels on the fly and improving their quality. In contrast to autoregressive formulations, such as the attention-based encoder-decoder and transducer, CTC is well suited for MPL, or PL-based semi-supervised ASR in general, owing to its simple/fast inference algorithm and robustness against generating collapsed labels. However, CTC generally yields inferior performance than the autoregressive models due to the conditional independence assumption, thereby limiting the performance of MPL. We propose to enhance MPL by introducing intermediate loss, inspired by the recent advances in CTC-based modeling. Specifically, we focus on self-conditional and hierarchical conditional CTC, that apply auxiliary CTC losses to intermediate layers such that the conditional independence assumption is explicitly relaxed. We also explore how pseudo-labels should be generated and used as supervision for intermediate losses. Experimental results in different semi-supervised settings demonstrate that the proposed approach outperforms MPL and improves an ASR model by up to a 12.1% absolute performance gain. In addition, our detailed analysis validates the importance of the intermediate loss.Comment: Submitted to ICASSP202

    BECTRA: Transducer-based End-to-End ASR with BERT-Enhanced Encoder

    Full text link
    We present BERT-CTC-Transducer (BECTRA), a novel end-to-end automatic speech recognition (E2E-ASR) model formulated by the transducer with a BERT-enhanced encoder. Integrating a large-scale pre-trained language model (LM) into E2E-ASR has been actively studied, aiming to utilize versatile linguistic knowledge for generating accurate text. One crucial factor that makes this integration challenging lies in the vocabulary mismatch; the vocabulary constructed for a pre-trained LM is generally too large for E2E-ASR training and is likely to have a mismatch against a target ASR domain. To overcome such an issue, we propose BECTRA, an extended version of our previous BERT-CTC, that realizes BERT-based E2E-ASR using a vocabulary of interest. BECTRA is a transducer-based model, which adopts BERT-CTC for its encoder and trains an ASR-specific decoder using a vocabulary suitable for a target task. With the combination of the transducer and BERT-CTC, we also propose a novel inference algorithm for taking advantage of both autoregressive and non-autoregressive decoding. Experimental results on several ASR tasks, varying in amounts of data, speaking styles, and languages, demonstrate that BECTRA outperforms BERT-CTC by effectively dealing with the vocabulary mismatch while exploiting BERT knowledge.Comment: Submitted to ICASSP202

    Segment-Level Vectorized Beam Search Based on Partially Autoregressive Inference

    Full text link
    Attention-based encoder-decoder models with autoregressive (AR) decoding have proven to be the dominant approach for automatic speech recognition (ASR) due to their superior accuracy. However, they often suffer from slow inference. This is primarily attributed to the incremental calculation of the decoder. This work proposes a partially AR framework, which employs segment-level vectorized beam search for improving the inference speed of an ASR model based on the hybrid connectionist temporal classification (CTC) attention-based architecture. It first generates an initial hypothesis using greedy CTC decoding, identifying low-confidence tokens based on their output probabilities. We then utilize the decoder to perform segment-level vectorized beam search on these tokens, re-predicting in parallel with minimal decoder calculations. Experimental results show that our method is 12 to 13 times faster in inference on the LibriSpeech corpus over AR decoding whilst preserving high accuracy.Comment: Accepted at ASRU 202
    corecore