15 research outputs found

    Data-driven Auditory Contrast Enhancement for Everyday Sounds and Sonifications

    Get PDF
    Hermann T, Weger M. Data-driven Auditory Contrast Enhancement for Everyday Sounds and Sonifications. In: Proceedings of the 25th International Conference on Auditory Display (ICAD 2019). Newcastle: ICAD; 2019.We introduce Auditory Contrast Enhancement (ACE) as a technique to enhance sounds at hand of a given collection of sound or sonification examples that belong to different classes, such as sounds of machines with and without a certain malfunction, or medical data sonifications for different pathologies/conditions. A frequent use case in inductive data mining is the discovery of patterns in which such groups can be discerned, to guide subsequent paths for modelling and feature extraction. ACE provides researchers with a set of methods to render focused auditory perspectives that accentuate inter-group differences and in turn also enhance the intra-group similarity, i.e, it warps sounds so that our human built-in metrics for assessing differences between sounds is better aligned to systematic differences between sounds belonging to different classes. We unfold and detail the concept along three different lines: temporal, spectral and spectrotemporal auditory contrast enhancement and we demonstrate their performance at hand of given sound and sonification collections

    AltAR/table: A Platform for Plausible Auditory Augmentation

    Get PDF
    Presented at the 27th International Conference on Auditory Display (ICAD 2022) 24-27 June 2022, Virtual conference.Auditory feedback from everyday interactions can be augmented to project digital information in the physical world. For that purpose, auditory augmentation modulates irrelevant aspects of already existing sounds while at the same time preserving relevant ones. A strategy for maintaining a certain level of plausibility is to metaphorically modulate the physical object itself. By mapping information to physical parameters instead of arbitrary sound parameters, it is assumed that even untrained users can draw on prior knowledge. Here we present AltAR/table, a hard- and software platform for plausible auditory augmentation of flat surfaces. It renders accurate augmentations of rectangular plates by capturing the structure-borne sound, feeding it through a physical sound model, and playing it back through the same object in real time. The implementation solves basic problems of equalization, active feedback control, spatialization, hand tracking, and low-latency signal processing. AltAR/table provides the technical foundations of object-centered auditory augmentations, for embedding sonifications into everyday objects such as tables, walls, or floors

    Plausible Auditory Augmentation of Physical Interaction

    Get PDF
    Weger M, Hermann T, Höldrich R. Plausible Auditory Augmentation of Physical Interaction. In: Proceedings of the 24th International Conference on Auditory Display. Sonification as ADSR. (ICAD 2018). Michigan: ICAD; 2018: 97-104.Interactions with physical objects usually evoke sounds, i.e., audi-tory feedback that depends on the interacting objects (e.g., table,hand, or pencil) and interaction type (e.g., tapping or scratching).The continuous real-time adaptation of sound during interactionenables the manipulation/refinement of perceived characteristics(size, material) of physical objects. Furthermore, when controlledby unrelated external data, the resulting ambient sonifications cankeep users aware of changing data. This article introduces the con-cept ofplausibilityto the topic of auditory augmentations of phys-ical interactions, aiming at providing an experimentation platformfor investigating surface-based physical interactions, understand-ing relevant acoustic cues, redefining these via auditory augmenta-tion / blended sonification and particularly to empirically measurethe plausibility limits of such auditory augmentations. Besidesconceptual contributions along the trade-off between plausibilityand usability, a practical experimentation system is introduced, to-gether with a very first qualitative pilot study

    Real-time Auditory Contrast Enhancement

    Get PDF
    Weger M, Hermann T, Höldrich R. Real-time Auditory Contrast Enhancement. In: Proceedings of the 25th International Conference on Auditory Display (ICAD 2019). Newcastle: International Community for Auditory Display (ICAD); In Press.Every day, we rely on the information that is encoded in the auditory feedback of our physical interactions. With the goal to perceptually enhance those sound characteristics that are relevant to us -- especially within professional practices such as percussion and auscultation, we introduce the method of real-time Auditory Contrast Enhancement (ACE). The method is derived from algorithms for speech enhancement as well from the remarkable sound processing mechanisms of our ears. ACE is achieved by individual sharpening of spectral as well as temporal structures contained in a sound. It is designed for real-time application with low latency. The discussed examples illustrate that the proposed algorithms are able to significantly enhance spectral and temporal contrast. ### Sound examples and interaction examples #### **ACE of Impact Sounds** ##### **S1.0-xxx.wav** see article for further details

    Data-driven auditory contrast enhancement for everyday sounds and sonifications

    Get PDF
    Presented at the 25th International Conference on Auditory Display (ICAD 2019) 23-27 June 2019, Northumbria University, Newcastle upon Tyne, UK.We introduce Auditory Contrast Enhancement (ACE) as a technique to enhance sounds at hand of a given collection of sound or sonification examples that belong to different classes, such as sounds of machines with and without a certain malfunction, or medical data sonifications for different pathologies/conditions. A frequent use case in inductive data mining is the discovery of patterns in which such groups can be discerned, to guide subsequent paths for modelling and feature extraction. ACE provides researchers with a set of methods to render focussed auditory perspectives that accentuate inter-group differences and in turn also enhance the intra-group similarity, i.e. it warps sounds so that our human built-in metrics for assessing differences between sounds is better aligned to systematic differences between sounds belonging to different classes. We unfold and detail the concept along three different lines: temporal, spectral and spectrotemporal auditory contrast enhancement and we demonstrate their performance at hand of given sound and sonification collections

    Recognizability and perceived urgency of bicycle bells

    Get PDF
    Raising awareness about how alarm sounds are perceived and evaluated by an individual in traffic scenery is important for developing new alarm designs, as well as for improving existing ones. Bearing a positive contribution to road safety, cyclists and pedestrians especially can benefit from appropriate alarming bell and horn sounds. Primarily, the alarm signal should evoke a precise idea of what is the source of the warning and the desired reaction to it. Furthermore, it should not be masked by other noises thus going undetected by the ear. Finally, an appropriate warning signal should transmit the urgency of a given situation, while at the same time, it should not cause other road users and pedestrians to startle. In two listening experiments, we examined the perception of commonly available bicycle bells and horns. Average typicality or recognizability as a bicycle bell among other everyday sounds has been investigated through a free identification task. In a second experiment, we tested perceived urgency of the warning sounds in relation to traffic noise. This article further provides a survey on non-verbal alarm design, as well as an analysis of acoustic properties of common bicycle bells and horns. Consequently, a linear regression model presents the relationship between named properties and perceived urgency. It is our intention to give an insight into the often unattended but important issue of the perception of auditory warning sounds in our everyday acoustic environment

    Plausible auditory augmentation of physical interaction

    Get PDF
    Interactions with physical objects usually evoke sounds, i.e., auditory feedback that depends on the interacting objects (e.g., table, hand, or pencil) and interaction type (e.g., tapping or scratching). The continuous real-time adaptation of sound during interaction enables the manipulation/refinement of perceived characteristics (size, material) of physical objects. Furthermore, when controlled by unrelated external data, the resulting ambient sonifications can keep users aware of changing data. This article introduces the concept of plausibility to the topic of auditory augmentations of physical interactions, aiming at providing an experimentation platform for investigating surface-based physical interactions, understanding relevant acoustic cues, redefining these via auditory augmentation / blended sonification and particularly to empirically measure the plausibility limits of such auditory augmentations. Besides conceptual contributions along the trade-off between plausibility and usability, a practical experimentation system is introduced, together with a very first qualitative pilot study

    Real-time auditory contrast enhancement

    Get PDF
    Presented at the 25th International Conference on Auditory Display (ICAD 2019) 23-27 June 2019, Northumbria University, Newcastle upon Tyne, UK.Every day, we rely on the information that is encoded in the auditory feedback of our physical interactions. With the goal to perceptually enhance those sound characteristics that are relevant to us - especially within professional practices such as percussion and auscultation - we introduce the method of real-time Auditory Contrast Enhancement (ACE). It is derived from algorithms for speech enhancement as well as from the remarkable sound processing mechanisms of our ears. ACE is achieved by individual sharpening of spectral and temporal structures contained in a sound while maintaining its natural gestalt. With regard to the targeted real-time applications, the proposed method is designed for low latency. As the discussed examples illustrate, it is able to significantly enhance spectral and temporal contrast

    Supplementary Material for "Data-driven Auditory Contrast Enhancement for Everyday Sounds and Sonifications"

    No full text
    Hermann T, Weger M. Supplementary Material for "Data-driven Auditory Contrast Enhancement for Everyday Sounds and Sonifications". Bielefeld University; 2019.We introduce Auditory Contrast Enhancement (ACE) as a technique to enhance sounds at hand of a given collection of sound or sonification examples that belong to different classes, such as sounds of machines with and without a certain malfunction, or medical data sonifications for different pathologies/conditions. A frequent use case in inductive data mining is the discovery of patterns in which such groups can be discerned, to guide subsequent paths for modelling and feature extraction. ACE provides researchers with a set of methods to render focused auditory perspectives that accentuate inter-group differences and in turn also enhance the intra-group similarity, i.e, it warps sounds so that our human built-in metrics for assessing differences between sounds is better aligned to systematic differences between sounds belonging to different classes. We unfold and detail the concept along three different lines: temporal, spectral and spectrotemporal auditory contrast enhancement and we demonstrate their performance at hand of given sound and sonification collections. #### **ACE of Impact Sounds: table interaction sounds** ##### **S1.0-table-stimuli.wav** see article for further details. ##### **S1.1-table__spectral__mode=median_filtered_ts__order=2.5__medfilt=9__nlgain=0.3.wav** see article for further details. ##### **S1.2-table__spectral__mode=tanh_gained_ts__order=6.25__medfilt=25__nlgain=0.16.wav** see article for further details. ##### **S1.3-table__spectral__mode=tanh_gained_ts__order=26.75__medfilt=17__nlgain=0.17.wav** see article for further details. ##### **S1.4-table__spectral__mode=db_median_filtered_ps__order=47.5__medfilt=25__nlgain=0.3.wav** see article for further details. ##### **S1.5-table__temporal__mode=median_filtered_ts__order=3.5__medfilt=5__nlgain=0.3.wav** see article for further details. ##### **S1.6-table__spectrotemporal__mode=only_highest_ts__ts_threshold=3.311__sigma=0.316__mix=1.0.wav** see article for further details. ##### **S1.7-table__spectrotemporal__mode=only_highest_ts__ts_threshold=13.18__sigma=0.316__mix=1.0.wav** see article for further details. ##### **S1.8-table__spectrotemporal__mode=blurred_ts_mask__ts_threshold=6.0256__sigma=0.331__mix=1.0.wav** see article for further details. ##### **S1.9-table__spectrotemporal__mode=blurred_ts_mask__ts_threshold=6.0256__sigma=0.871__mix=0.998.wav** see article for further details. #### **ACE of Impact Sounds: finger snap interaction sounds** ##### **S2.0-snap-stimuli.wav** see article for further details. ##### **S2.1-snap__spectral__mode=median_filtered_ts__order=1.0__medfilt=1__nlgain=0.3.wav** see article for further details. ##### **S2.2-snap__spectral__mode=tanh_gained_ts__order=4.5__medfilt=1__nlgain=0.3.wav** see article for further details. ##### **S2.3-snap__spectral__mode=tanh_gained_ts__order=4.5__medfilt=1__nlgain=0.12.wav** see article for further details. ##### **S2.4-snap__spectral__mode=median_filtered_ts__order=4.5__medfilt=1__nlgain=0.3.wav** see article for further details. ##### **S2.5-snap__spectrotemporal__mode=only_highest_ts__ts_threshold=3.02__sigma=0.316__mix=1.0.wav** see article for further details. #### **ACE of synthesized sounds: SuperCollider Klank plus Noise** ##### **S3.0-klanks-stimuli.wav** see article for further details. ##### **S3.1-klanks__spectral__mode=median_filtered_ts__medfilt=15__order=1.75__nlgain=0.0.wav** see article for further details. ##### **S3.2-klanks__spectral__mode=tanh_gained_ts__order=3.0__medfilt=9__nlgain=0.21.wav** see article for further details. ##### **S3.3-klanks__spectral__mode=tanh_gained_ts__medfilt=27__order=7.0__nlgain=0.23.wav** see article for further details. ##### **S3.4-klanks__spectral__mode=median_filtered_ts__order=9.0__medfilt=1__nlgain=0.04.wav** see article for further details. ##### **S3.5-klanks__spectrotemporal__mode=blurred_ts_mask__ts_threshold=0.832__sigma=0.661__mix=0.998.wav** see article for further details. #### **ACE of continuous sounds: Vowel sounds class a vs. ä** ##### **S4.0-vowel-stimuli.wav** see article for further details. ##### **S4.1-vowel__spectral__mode=tanh_gained_ts__order=4.25__medfilt=39__nlgain=0.09.wav** see article for further details. ##### **S4.2-vowel__spectral__mode=median_filtered_ts__order=10.75__medfilt=39__nlgain=0.01.wav** see article for further details. ##### **S4.3-vowel__spectrotemporal__mode=only_highest_ts__ts_threshold=2.291__sigma=0.912__mix=1.0.wav** see article for further details. ##### **S4.4-vowel-a-ae-transition-orig.wav** see article for further details. ##### **S4.5-vowel-a-ae-transition-aced.wav** see article for further details. #### **ACE of time series sonifications: daily energy consumption profiles of a building** ##### **S5.0-building-stimuli.wav** see article for further details. ##### **S5.1-building__spectral__mode=median_filtered_ts__medfilt=5__order=1.0__nlgain=0.01.wav** see article for further details. ##### **S5.2-building__spectral__mode=tanh_gained_ts__medfilt=3__order=0.5__nlgain=0.01.wav** see article for further details. ##### **S5.3-building__temporal__mode=median_filtered_ts__medfilt=7__order=2.75__nlgain=0.3.wav** see article for further details. ##### **S5.4-building__spectrotemporal__mode=only_highest_ts__ts_threshold=3.981__sigma=0.603__mix=0.995.wav** see article for further details. #### **ACE of manipulated music recordings: Temporal ACE gives condensed difference thumbnails** ##### **S6.0-miles-stimuli.wav** see article for further details. ##### **S6.1-miles__temporal__mode=median_filtered_ts__order=0.5__medfilt=7__nlgain=0.3.wav** see article for further details. ##### **S6.2-miles__temporal__mode=tanh_gained_ts__order=0.5__medfilt=7__nlgain=0.12.wav** see article for further details

    Supplementary Material for "Real-time Auditory Contrast Enhancement"

    No full text
    Weger M, Hermann T, Höldrich R. Supplementary Material for "Real-time Auditory Contrast Enhancement". Bielefeld University; 2019.Every day, we rely on the information that is encoded in the auditory feedback of our physical interactions. With the goal to perceptually enhance those sound characteristics that are relevant to us – especially within professional practices such as percussion and auscultation – we introduce the method of real-time Audi- tory Contrast Enhancement (ACE). It is derived from algorithms for speech enhancement as well as from the remarkable sound processing mechanisms of our ears. ACE is achieved by individual sharpening of spectral and temporal structures contained in a sound while maintaining its natural gestalt. With regard to the targeted real-time applications, the proposed method is designed for low latency. As the discussed examples illustrate, it is able to significantly enhance spectral and temporal contrast. ### Sound examples #### **ACE of impact Sounds** ##### **S1.1_knock-knock.wav** see article for further details. ##### **S1.2_knock-knock_LI.wav** see article for further details. ##### **S1.3_knock-knock_LI+EX.wav** see article for further details. ##### **S1.4_knock-knock_LI+EX+DP.wav** see article for further details. ##### **S1.5_knock-knock_TCE.wav** see article for further details. ##### **S1.6_knock-knock_LI+TCE.wav** see article for further details. ##### **S1.7_knock-knock_LI+EX+TCE.wav** see article for further details. ##### **S1.8_knock-knock_LI+EX+DP+TCE.wav** see article for further details. #### **ACE of machine sounds** ##### **S2.1_machine.wav** see article for further details. ##### **S2.2_machine_EX.wav** see article for further details. ##### **S2.3_machine_TCE.wav** see article for further details. ##### **S2.4_machine_EX+TCE.wav** see article for further details
    corecore