18,554 research outputs found
Application of the wavelet transform for speech processing
Speaker identification and word spotting will shortly play a key role in space applications. An approach based on the wavelet transform is presented that, in the context of the 'modulation model,' enables extraction of speech features which are used as input for the classification process
Image compression using discrete cosine transform and wavelet transform and performance comparison
Image compression deals with reducing the size of image which is performed with the help of transforms. In this project we have taken the Input image and applied wavelet techniques for image compression and have compared the result with the popular DCT image compression. WT provided better result as far as properties like RMS error, image intensity and execution time is concerned. Now a days wavelet theory based technique has emerged in different
signal and image processing application including speech, image processing and computer vision. In particular Wavelet Transform is of interest for the analysis of non-stationary signals. In the WT at high frequencies short windows and at low frequencies long windows are used. Since discrete wavelet is essentially sub band–coding system, sub band coders have been quit successful in speech and image compression. It is clear that DWT has potential application in compression problem
A Wavelet Transform Module for a Speech Recognition Virtual Machine
This work explores the trade-offs between time and frequency information during the feature extraction process of an automatic speech recognition (ASR) system using wavelet transform (WT) features instead of Mel-frequency cepstral coefficients (MFCCs) and the benefits of combining the WTs and the MFCCs as inputs to an ASR system. A virtual machine from the Speech Recognition Virtual Kitchen resource (www.speechkitchen.org) is used as the context for implementing a wavelet signal processing module in a speech recognition system. Contributions include a comparison of MFCCs and WT features on small and large vocabulary tasks, application of combined MFCC and WT features on a noisy environment task, and the implementation of an expanded signal processing module in an existing recognition system. The updated virtual machine, which allows straightforward comparisons of signal processing approaches, is available for research and education purposes
Isolated Word Speech Recognition Using Mixed Transform
Methods of speech recognition have been the subject of several studies over the past decade. Speech recognition has been one of the most exciting areas of the signal processing. Mixed transform is a useful tool for speech signal processing; it is developed for its abilities of improvement in feature extraction. Speech recognition includes three important stages, preprocessing, feature extraction, and classification. Recognition accuracy is so affected by the features extraction stage; therefore different models of mixed transform for feature extraction were proposed. The properties of the recorded isolated word will be 1-D, which achieve the conversion of each 1-D word into a 2-D form. The second step of the word recognizer requires, the application of 2-D FFT, Radon transform, the 1-D IFFT,and 1-D discrete wavelet transforms were used in the first proposed model, while discrete multicircularlet transform was used in the second proposed model. The final stage of the proposed models includes the use of the dynamic time warping algorithm for recognition tasks. The performance of the proposed systems was evaluated using forty different isolated Arabic words that are recorded fifteen times in a studio for speaker dependant. The result shows recognition accuracy of (91% and 89%) using discrete wavelet transform type Daubechies (Db1) and (Db4) respectively, and the accuracy score between (87%-93%) was achieved using
discrete multicircularlet transform for 9 sub bands
Spectral analysis for nonstationary audio
A new approach for the analysis of nonstationary signals is proposed, with a
focus on audio applications. Following earlier contributions, nonstationarity
is modeled via stationarity-breaking operators acting on Gaussian stationary
random signals. The focus is on time warping and amplitude modulation, and an
approximate maximum-likelihood approach based on suitable approximations in the
wavelet transform domain is developed. This paper provides theoretical analysis
of the approximations, and introduces JEFAS, a corresponding estimation
algorithm. The latter is tested and validated on synthetic as well as real
audio signal.Comment: IEEE/ACM Transactions on Audio, Speech and Language Processing,
Institute of Electrical and Electronics Engineers, In pres
- …