4 research outputs found
CIF-PT: Bridging Speech and Text Representations for Spoken Language Understanding via Continuous Integrate-and-Fire Pre-Training
Speech or text representation generated by pre-trained models contains
modal-specific information that could be combined for benefiting spoken
language understanding (SLU) tasks. In this work, we propose a novel
pre-training paradigm termed Continuous Integrate-and-Fire Pre-Training
(CIF-PT). It relies on a simple but effective frame-to-token alignment:
continuous integrate-and-fire (CIF) to bridge the representations between
speech and text. It jointly performs speech-to-text training and language model
distillation through CIF as the pre-training (PT). Evaluated on SLU benchmark
SLURP dataset, CIF-PT outperforms the state-of-the-art model by 1.94% of
accuracy and 2.71% of SLU-F1 on the tasks of intent classification and slot
filling, respectively. We also observe the cross-modal representation extracted
by CIF-PT obtains better performance than other neural interfaces for the tasks
of SLU, including the dominant speech representation learned from
self-supervised pre-training.Comment: Accepted by ACL 2023 Finding
Robust estimation of bacterial cell count from optical density
Optical density (OD) is widely used to estimate the density of cells in liquid culture, but cannot be compared between instruments without a standardized calibration protocol and is challenging to relate to actual cell count. We address this with an interlaboratory study comparing three simple, low-cost, and highly accessible OD calibration protocols across 244 laboratories, applied to eight strains of constitutive GFP-expressing E. coli. Based on our results, we recommend calibrating OD to estimated cell count using serial dilution of silica microspheres, which produces highly precise calibration (95.5% of residuals <1.2-fold), is easily assessed for quality control, also assesses instrument effective linear range, and can be combined with fluorescence calibration to obtain units of Molecules of Equivalent Fluorescein (MEFL) per cell, allowing direct comparison and data fusion with flow cytometry measurements: in our study, fluorescence per cell measurements showed only a 1.07-fold mean difference between plate reader and flow cytometry data
Improving Contextual Representation with Gloss Regularized Pre-training
Though achieving impressive results on many NLP tasks, the BERT-like masked
language models (MLM) encounter the discrepancy between pre-training and
inference. In light of this gap, we investigate the contextual representation
of pre-training and inference from the perspective of word probability
distribution. We discover that BERT risks neglecting the contextual word
similarity in pre-training. To tackle this issue, we propose an auxiliary gloss
regularizer module to BERT pre-training (GR-BERT), to enhance word semantic
similarity. By predicting masked words and aligning contextual embeddings to
corresponding glosses simultaneously, the word similarity can be explicitly
modeled. We design two architectures for GR-BERT and evaluate our model in
downstream tasks. Experimental results show that the gloss regularizer benefits
BERT in word-level and sentence-level semantic representation. The GR-BERT
achieves new state-of-the-art in lexical substitution task and greatly promotes
BERT sentence representation in both unsupervised and supervised STS tasks.Comment: Accepted to Findings of NAACL 202