1,677 research outputs found

    Variance of spectral entropy (VSE): an SNR estimator for speech enhancement in hearing aids

    Get PDF
    In everyday situations an individual can encounter a variety of acoustic environments. For an individual with a hearing aid following speech in different types of background noise can often present a challenge. For this reason, estimating the signal-to-noise ratio (SNR) is a key factor to consider in hearing-aid design. The ability to adjust a noise reduction algorithm according to the SNR could provide the flexibility required to improve speech intelligibility in varying levels of background noise. However, most of the current high-accuracy SNR estimation methods are relatively complex and may inhibit the performance of hearing aids. This study investigates the advantages of incorporating a spectral entropy method to estimate SNR for speech enhancement in hearing aids; in particular a variance of spectral entropy (VSE) measure. The VSE approach avoids some of the complex computational steps of traditional statistical-model based SNR estimation methods by only measuring the spectral entropy among frequency channels of interest within the hearing aid. For this study, the SNR was estimated using the spectral entropy method in different types of noise. The variance of the spectral entropy in a hearing-aid model with 10 peripheral frequency channels was used to measure the SNR. By measuring the variance of the spectral entropy at input SNR levels between -10 dB to 20 dB, the relationship function between the SNR and the VSE was estimated. The VSE for the speech-in-noise was measured at temporal intervals of 1.5s. The VSE method demonstrates a more reliable performance in different types of background noise, in particular for low-number of speakers babble noise when compared to the US National Institute of Standards and Technology (NIST) or Waveform Amplitude Distribution Analysis (WADA) methods. The VSE method may also reduce additional computational steps (reducing system delays) making it more appropriate for implementation in hearing aids where system delays should be minimized as much as possible

    A Computation Efficient Voice Activity Detector for Low Signal-to-Noise Ratio in Hearing Aids

    Get PDF
    This paper proposes a spectral entropy-based voice activity detection method, which is computationally efficient for hearing aids. The method is highly accurate at low SNR levels by using the spectral entropy which is more robust against changes of the noise power. Compared with the traditional fast Fourier transform based spectral entropy approaches, the proposed method of calculating the spectral entropy using the outputs of a hearing aid filter-bank significantly reduces the computational complexity. The performance of the proposed method was evaluated and compared with two other computationally efficient methods. At negative SNR levels, the proposed method has an accuracy of more than 5% higher than the power-based method with the number of floating-point operations only about 1/100 of that of the statistical model based method

    Deep learning-based denoising streamed from mobile phones improves speech-in-noise understanding for hearing aid users

    Full text link
    The hearing loss of almost half a billion people is commonly treated with hearing aids. However, current hearing aids often do not work well in real-world noisy environments. We present a deep learning based denoising system that runs in real time on iPhone 7 and Samsung Galaxy S10 (25ms algorithmic latency). The denoised audio is streamed to the hearing aid, resulting in a total delay of around 75ms. In tests with hearing aid users having moderate to severe hearing loss, our denoising system improves audio across three tests: 1) listening for subjective audio ratings, 2) listening for objective speech intelligibility, and 3) live conversations in a noisy environment for subjective ratings. Subjective ratings increase by more than 40%, for both the listening test and the live conversation compared to a fitted hearing aid as a baseline. Speech reception thresholds, measuring speech understanding in noise, improve by 1.6 dB SRT. Ours is the first denoising system that is implemented on a mobile device, streamed directly to users' hearing aids using only a single channel as audio input while improving user satisfaction on all tested aspects, including speech intelligibility. This includes overall preference of the denoised and streamed signal over the hearing aid, thereby accepting the higher latency for the significant improvement in speech understanding

    Automatic User Preferences Selection of Smart Hearing Aid Using BioAid

    Get PDF
    Noisy environments, changes and variations in the volume of speech, and non-face-to-face conversations impair the user experience with hearing aids. Generally, a hearing aid amplifies sounds so that a hearing-impaired person can listen, converse, and actively engage in daily activities. Presently, there are some sophisticated hearing aid algorithms available that operate on numerous frequency bands to not only amplify but also provide tuning and noise filtering to minimize background distractions. One of those is the BioAid assistive hearing system, which is an open-source, freely available downloadable app with twenty-four tuning settings. Critically, with this device, a person suffering with hearing loss must manually alter the settings/tuning of their hearing device when their surroundings and scene changes in order to attain a comfortable level of hearing. However, this manual switching among multiple tuning settings is inconvenient and cumbersome since the user is forced to switch to the state that best matches the scene every time the auditory environment changes. The goal of this study is to eliminate this manual switching and automate the BioAid with a scene classification algorithm so that the system automatically identifies the user-selected preferences based on adequate training. The aim of acoustic scene classification is to recognize the audio signature of one of the predefined scene classes that best represent the environment in which it was recorded. BioAid, an open-source biological inspired hearing aid algorithm, is used after conversion to Python. The proposed method consists of two main parts: classification of auditory scenes and selection of hearing aid tuning settings based on user experiences. The DCASE2017 dataset is utilized for scene classification. Among the many classifiers that were trained and tested, random forests have the highest accuracy of 99.7%. In the second part, clean speech audios from the LJ speech dataset are combined with scenes, and the user is asked to listen to the resulting audios and adjust the presets and subsets. A CSV file stores the selection of presets and subsets at which the user can hear clearly against the scenes. Various classifiers are trained on the dataset of user preferences. After training, clean speech audio was convolved with the scene and fed as input to the scene classifier that predicts the scene. The predicted scene was then fed as input to the preset classifier that predicts the user’s choice for preset and subset. The BioAid is automatically tuned to the predicted selection. The accuracy of random forest in the prediction of presets and subsets was 100%. This proposed approach has great potential to eliminate the tedious manual switching of hearing assistive device parameters by allowing hearing-impaired individuals to actively participate in daily life by automatically adjusting hearing aid settings based on the acoustic scen

    How does auditory training work? Joined up thinking and listening

    Get PDF
    Auditory training aims to compensate for degradation in the auditory signal and is offered as an intervention to help alleviate the most common complaint in people with hearing loss, understanding speech in a background noise. Yet there remain many unanswered questions. This article reviews some of the key pieces of evidence that assess the evidence for whether, and how, auditory training benefits adults with hearing loss. The evidence supports that improvements occur on the trained task; however, transfer of that learning to generalized real-world benefit is much less robust. For more than a decade, there has been an increasing awareness of the role that cognition plays in listening. But more recently in the auditory training literature, there has been an increased focus on assessing how cognitive performance relevant for listening may improve with training. We argue that this is specifically the case for measures that index executive processes, such as monitoring, attention switching, and updating of working memory, all of which are required for successful listening and communication in challenging or adverse listening conditions. We propose combined auditory-cognitive training approaches, where training interventions develop cognition embedded within auditory tasks, which are most likely to offer generalized benefits to the real-world listening abilities of people with hearing loss

    An objective evaluation of Hearing Aids and DNN-based speech enhancement in complex acoustic scenes

    Full text link
    We investigate the objective performance of five high-end commercially available Hearing Aid (HA) devices compared to DNN-based speech enhancement algorithms in complex acoustic environments. To this end, we measure the HRTFs of a single HA device to synthesize a binaural dataset for training two state-of-the-art causal and non-causal DNN enhancement models. We then generate an evaluation set of realistic speech-in-noise situations using an Ambisonics loudspeaker setup and record with a KU100 dummy head wearing each of the HA devices, both with and without the conventional HA algorithms, applying the DNN enhancers to the latter. We find that the DNN-based enhancement outperforms the HA algorithms in terms of noise suppression and objective intelligibility metrics.Comment: Accepted to WASPAA2

    Reconfigurable Multiband Dynamic Range Compression-based FRM Filter for Hearing Aid

    Get PDF
    In this research, we present an innovative method for enhancing the performance of hearing aids using a Multiband Dynamic Range Compression-based Reconfigurable Frequency Response Masking (FRM) Filterbank. First, a unform16-band reconfigurable filter bank, which is reconfigurable, is designed utilizing the FRM scheme. The strategic arrangement of each sub-band within the proposed filter bank is meticulously prepared to optimize the matching performance. Based on the hearing characteristics of patients, the sub-bands can be distributed in low, medium, and high-frequency regions. Also, the gain can be adjusted per the patient's hearing profile from their audiogram for better auditory compensation. Further, the Multiband Dynamic Range Compression (MBDRC) technique is applied to address the specific needs of individuals with different frequency-dependent hearing impairments. It involves using dynamic range compression independently to different frequency sub-bands within a filter bank. In MBDRC, the compression parameters, such as compression threshold and ratio, can be adjusted independently for every subband. It allows for a more tailored approach to address the specific hearing needs of different frequency regions. If an individual has more severe hearing loss in high-frequency regions, higher compression ratios and lower compression thresholds can be applied to those subbands to amplify and improve audibility for high-frequency sounds. Once dynamic range compression is applied to each sub-band, the resultant sub-bands are reassembled to yield the ultimate output signal, which can subsequently be transmitted to the speaker or receiver of the hearing aid. A GUI can be helpful for better visualization and parameter control, including gain adjustment and compression parameters of this entire process. With this aim in mind, a GUI has been developed on MATLAB. Different audio files can be imported, and their frequency response can be generated and observed. Based on a person's audiogram, the control parameters can be set to low, medium, or high. Their sub-band distribution in low, medium, and high-frequency regions can be visualized. Further, the filter bank makes automatic gain adjustments, as seen in the GUI. The gain points for each band can also be manually adjusted according to users' hearing characteristics to minimize the error. Also, the compression parameters can be set separately for each subband as per the hearing requirement of the patient. Further, the processed output can be visualized in the output frequency response tab, and the input and output audio signals can be analyzed

    Challenges in Developing Applications for Aging Populations

    Get PDF
    Elderly individuals can greatly benefit from the use of computer applications, which can assist in monitoring health conditions, staying in contact with friends and family, and even learning new things. However, developing accessible applications for an elderly user can be a daunting task for developers. Since the advent of the personal computer, the benefits and challenges of developing applications for older adults have been a hot topic of discussion. In this chapter, the authors discuss the various challenges developers who wish to create applications for the elderly computer user face, including age-related impairments, generational differences in computer use, and the hardware constraints mobile devices pose for application developers. Although these challenges are concerning, each can be overcome after being properly identified
    • …
    corecore