11 research outputs found
Automatic correction of disfluent spoken queries
A user’s interaction with a virtual assistant typically involves spoken requests, queries, and commands which often includes disfluencies. This disclosure describes techniques to automatically correct disfluent queries. Per techniques of this disclosure, a disfluency correction machine learning model is utilized to convert a disfluent query to a corresponding fluent query. Lexical features extracted from the disfluent query are utilized to determine a portion of the query that is removed from the disfluent query to convert it to a fluent query. The model is trained using pairs of queries
Multi-Task Self-Supervised Learning for Disfluency Detection
Most existing approaches to disfluency detection heavily rely on
human-annotated data, which is expensive to obtain in practice. To tackle the
training data bottleneck, we investigate methods for combining multiple
self-supervised tasks-i.e., supervised tasks where data can be collected
without manual labeling. First, we construct large-scale pseudo training data
by randomly adding or deleting words from unlabeled news data, and propose two
self-supervised pre-training tasks: (i) tagging task to detect the added noisy
words. (ii) sentence classification to distinguish original sentences from
grammatically-incorrect sentences. We then combine these two tasks to jointly
train a network. The pre-trained network is then fine-tuned using
human-annotated disfluency detection training data. Experimental results on the
commonly used English Switchboard test set show that our approach can achieve
competitive performance compared to the previous systems (trained using the
full dataset) by using less than 1% (1000 sentences) of the training data. Our
method trained on the full dataset significantly outperforms previous methods,
reducing the error by 21% on English Switchboard