Real-Time Integration of Dynamic Context Information for Improving Automatic Speech Recognition

Abstract

The use of prior situational/contextual knowledge about a given task can significantly improve automatic speech recognition (ASR) performance. This is typically done through adaptation of acoustic or language models if data is available or using knowledge-based rescoring. The main adaptation techniques, however, are either domain-specific, which makes them inadequate for other tasks, or static and offline, and therefore cannot deal with dynamic knowledge. To circumvent this problem, we propose a real-time system which dynamically integrates situational context into ASR. The context integration is done either post-recognition, in which case a weighted Levenshtein distance between the ASR hypotheses and the context information based on the ASR confidence scores is proposed to extract the most likely sequence of spoken words, or pre-recognition, where the search space is adjusted to the new situational knowledge through adaptation of the finite state machine modeling the spoken language. Experiments conducted on 3 hours of Air Traffic Control (ATC) data achieved a 51% reduction of the Command Error Rate (CmdER) which is used as evaluation metric in the ATC domain

Similar works

Full text

thumbnail-image

Institute of Transport Research:Publications

Full text is not available
oai:elib.dlr.de:96937Last time updated on 4/28/2016

Having an issue?

Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.