CORE
🇺🇦
make metadata, not war
Services
Services overview
Explore all CORE services
Access to raw data
API
Dataset
FastSync
Content discovery
Recommender
Discovery
OAI identifiers
OAI Resolver
Managing content
Dashboard
Bespoke contracts
Consultancy services
Support us
Support us
Membership
Sponsorship
Community governance
Advisory Board
Board of supporters
Research network
About
About us
Our mission
Team
Blog
FAQs
Contact us
Speech enhancement Algorithm based on super-Gaussian modeling and orthogonal polynomials
Authors
Sadiq H. Abdulhussain
Feras Al-Obeidat
+4 more
Thar Baker
Wissam A. Jassim
Basheera M. Mahmmod
Abd Rahman Ramli
Publication date
1 January 2019
Publisher
ZU Scholars
Doi
Cite
Abstract
© 2020 Lippincott Williams and Wilkins. All rights reserved. Different types of noise from the surrounding always interfere with speech and produce annoying signals for the human auditory system. To exchange speech information in a noisy environment, speech quality and intelligibility must be maintained, which is a challenging task. In most speech enhancement algorithms, the speech signal is characterized by Gaussian or super-Gaussian models, and noise is characterized by a Gaussian prior. However, these assumptions do not always hold in real-life situations, thereby negatively affecting the estimation, and eventually, the performance of the enhancement algorithm. Accordingly, this paper focuses on deriving an optimum low-distortion estimator with models that fit well with speech and noise data signals. This estimator provides minimum levels of speech distortion and residual noise with additional improvements in speech perceptual aspects via four key steps. First, a recent transform based on an orthogonal polynomial is used to transform the observation signal into a transform domain. Second, the noise classification based on feature extraction is adopted to find accurate and mutable models for noise signals. Third, two stages of nonlinear and linear estimators based on the minimum mean square error (MMSE) and new models for speech and noise are derived to estimate a clean speech signal. Finally, the estimated speech signal in the time domain is determined by considering the inverse of the orthogonal transform. The results show that the average classification accuracy of the proposed approach is 99.43%. In addition, the proposed algorithm significantly outperforms existing speech estimators in terms of quality and intelligibility measures
Similar works
Full text
Open in the Core reader
Download PDF
Available Versions
ZU Scholars (Zayed University)
See this paper in CORE
Go to the repository landing page
Download from data provider
oai:zuscholars.zu.ac.ae:works-...
Last time updated on 03/12/2021
University of Brighton Research Portal
See this paper in CORE
Go to the repository landing page
Download from data provider
oai:pure.atira.dk:publications...
Last time updated on 21/01/2023