988,552 research outputs found

    Representation of Sound Categories in Auditory Cortical Maps

    Full text link
    We used functional magnetic resonance imaging (fMRI) to investigate the representation of sound categories in human auditory cortex. Experiment 1 investigated the representation of prototypical and non-prototypical examples of a vowel sound. Listening to prototypical examples of a vowel resulted in less auditory cortical activation than listening to nonprototypical examples. Experiments 2 and 3 investigated the effects of categorization training and discrimination training with novel non-speech sounds on auditory cortical representations. The two training tasks were shown to have opposite effects on the auditory cortical representation of sounds experienced during training: discrimination training led to an increase in the amount of activation caused by the training stimuli, whereas categorization training led to decreased activation. These results indicate that the brain efficiently shifts neural resources away from regions of acoustic space where discrimination between sounds is not behaviorally important (e.g., near the center of a sound category) and toward regions where accurate discrimination is needed. The results also provide a straightforward neural account of learned aspects of categorical perception: sounds from the center of a category are more difficult to discriminate from each other than sounds near category boundaries because they are represented by fewer cells in the auditory cortical areas.National Institute on Deafness and Other Communication Disorders (R01 DC02852

    Ambient Sound Provides Supervision for Visual Learning

    Full text link
    The sound of crashing waves, the roar of fast-moving cars -- sound conveys important information about the objects in our surroundings. In this work, we show that ambient sounds can be used as a supervisory signal for learning visual models. To demonstrate this, we train a convolutional neural network to predict a statistical summary of the sound associated with a video frame. We show that, through this process, the network learns a representation that conveys information about objects and scenes. We evaluate this representation on several recognition tasks, finding that its performance is comparable to that of other state-of-the-art unsupervised learning methods. Finally, we show through visualizations that the network learns units that are selective to objects that are often associated with characteristic sounds.Comment: ECCV 201

    Acoustic Scene Classification by Implicitly Identifying Distinct Sound Events

    Full text link
    In this paper, we propose a new strategy for acoustic scene classification (ASC) , namely recognizing acoustic scenes through identifying distinct sound events. This differs from existing strategies, which focus on characterizing global acoustical distributions of audio or the temporal evolution of short-term audio features, without analysis down to the level of sound events. To identify distinct sound events for each scene, we formulate ASC in a multi-instance learning (MIL) framework, where each audio recording is mapped into a bag-of-instances representation. Here, instances can be seen as high-level representations for sound events inside a scene. We also propose a MIL neural networks model, which implicitly identifies distinct instances (i.e., sound events). Furthermore, we propose two specially designed modules that model the multi-temporal scale and multi-modal natures of the sound events respectively. The experiments were conducted on the official development set of the DCASE2018 Task1 Subtask B, and our best-performing model improves over the official baseline by 9.4% (68.3% vs 58.9%) in terms of classification accuracy. This study indicates that recognizing acoustic scenes by identifying distinct sound events is effective and paves the way for future studies that combine this strategy with previous ones.Comment: code URL typo, code is available at https://github.com/hackerekcah/distinct-events-asc.gi

    Representation, Interpretation, and Surrogative Reasoning

    Get PDF
    In this paper, I develop Mauricio Suárez’s distinction between denotation, epistemic representation, and faithful epistemic representation. I then outline an interpretational account of epistemic representation, according to which a vehicle represents a target for a certain user if and only if the user adopts an interpretation of the vehicle in terms of the target, which would allow them to perform valid (but not necessarily sound) surrogative inferences from the model to the system. The main difference between the interpretational conception I defend here and Suárez’s inferential conception is that the interpretational account is a substantial account—interpretation is not just a “symptom” of representation; it is that in virtue of what something is an epistemic representation of a something else

    Directivity patterns of laser-generated sound in solids: Effects of optical and thermal parameters

    Get PDF
    In the present paper, directivity patterns of laser-generated sound in solids are investigated theoretically. Two main approaches to the calculation of directivity patterns of laser-generated sound are discussed for the most important case of thermo-optical regime of generation. The first approach, which is widely used in practice, is based on the simple modelling of the equivalent thermo-optical source as a mechanical dipole comprising two horizontal forces applied to the surface in opposite directions. The second approach is based on the rigorous theory that takes into account all acoustical, optical and thermal parameters of a solid material and all geometrical and physical parameters of a laser beam. Directivity patterns of laser-generated bulk longitudinal and shear elastic waves, as well as the amplitudes of generated Rayleigh surface waves, are calculated for different values of physical and geometrical parameters and compared with the directivity patterns calculated in case of dipole-source representation. It is demonstrated that the simple approach using a dipole-source representation of laser-generated sound is rather limited, especially for description of generated longitudinal acoustic waves. A practical criterion is established to define the conditions under which the dipole-source representation gives predictions with acceptable errors. It is shown that, for radiation in the normal direction to the surface, the amplitudes of longitudinal waves are especially sensitive to the values of thermal parameters and of the acoustic reflection coefficient from a free solid surface. A discussion is given on the possibility of using such a high sensitivity to the values of the reflection coefficient for investigation of surface properties of real solids.Comment: 14 pages, 7 figure
    corecore