988,552 research outputs found
Representation of Sound Categories in Auditory Cortical Maps
We used functional magnetic resonance imaging (fMRI) to investigate the representation of sound categories in human auditory cortex. Experiment 1 investigated the representation of prototypical and non-prototypical examples of a vowel sound. Listening to prototypical examples of a vowel resulted in less auditory cortical activation than listening to nonprototypical examples. Experiments 2 and 3 investigated the effects of categorization training and discrimination training with novel non-speech sounds on auditory cortical representations. The two training tasks were shown to have opposite effects on the auditory cortical representation of sounds experienced during training: discrimination training led to an increase in the amount of activation caused by the training stimuli, whereas categorization training led to decreased activation. These results indicate that the brain efficiently shifts neural resources away from regions of acoustic space where discrimination between sounds is not behaviorally important (e.g., near the center of a sound category) and toward regions where accurate discrimination is needed. The results also provide a straightforward neural account of learned aspects of categorical perception: sounds from the center of a category are more difficult to discriminate from each other than sounds near category boundaries because they are represented by fewer cells in the auditory cortical areas.National Institute on Deafness and Other Communication Disorders (R01 DC02852
Ambient Sound Provides Supervision for Visual Learning
The sound of crashing waves, the roar of fast-moving cars -- sound conveys
important information about the objects in our surroundings. In this work, we
show that ambient sounds can be used as a supervisory signal for learning
visual models. To demonstrate this, we train a convolutional neural network to
predict a statistical summary of the sound associated with a video frame. We
show that, through this process, the network learns a representation that
conveys information about objects and scenes. We evaluate this representation
on several recognition tasks, finding that its performance is comparable to
that of other state-of-the-art unsupervised learning methods. Finally, we show
through visualizations that the network learns units that are selective to
objects that are often associated with characteristic sounds.Comment: ECCV 201
Acoustic Scene Classification by Implicitly Identifying Distinct Sound Events
In this paper, we propose a new strategy for acoustic scene classification
(ASC) , namely recognizing acoustic scenes through identifying distinct sound
events. This differs from existing strategies, which focus on characterizing
global acoustical distributions of audio or the temporal evolution of
short-term audio features, without analysis down to the level of sound events.
To identify distinct sound events for each scene, we formulate ASC in a
multi-instance learning (MIL) framework, where each audio recording is mapped
into a bag-of-instances representation. Here, instances can be seen as
high-level representations for sound events inside a scene. We also propose a
MIL neural networks model, which implicitly identifies distinct instances
(i.e., sound events). Furthermore, we propose two specially designed modules
that model the multi-temporal scale and multi-modal natures of the sound events
respectively. The experiments were conducted on the official development set of
the DCASE2018 Task1 Subtask B, and our best-performing model improves over the
official baseline by 9.4% (68.3% vs 58.9%) in terms of classification accuracy.
This study indicates that recognizing acoustic scenes by identifying distinct
sound events is effective and paves the way for future studies that combine
this strategy with previous ones.Comment: code URL typo, code is available at
https://github.com/hackerekcah/distinct-events-asc.gi
Representation, Interpretation, and Surrogative Reasoning
In this paper, I develop Mauricio Suárez’s distinction between denotation, epistemic representation, and faithful epistemic representation. I then outline an interpretational account of epistemic representation, according to which a vehicle represents a target for a certain user if and only if the user adopts an interpretation of the vehicle in terms of the target, which would allow them to perform valid (but not necessarily sound) surrogative inferences from the model to the system. The main difference between the interpretational conception I defend here and Suárez’s inferential conception is that the interpretational account is a substantial account—interpretation is not just a “symptom” of representation; it is that in virtue of what something is an epistemic representation of a something else
Directivity patterns of laser-generated sound in solids: Effects of optical and thermal parameters
In the present paper, directivity patterns of laser-generated sound in solids
are investigated theoretically. Two main approaches to the calculation of
directivity patterns of laser-generated sound are discussed for the most
important case of thermo-optical regime of generation. The first approach,
which is widely used in practice, is based on the simple modelling of the
equivalent thermo-optical source as a mechanical dipole comprising two
horizontal forces applied to the surface in opposite directions. The second
approach is based on the rigorous theory that takes into account all
acoustical, optical and thermal parameters of a solid material and all
geometrical and physical parameters of a laser beam. Directivity patterns of
laser-generated bulk longitudinal and shear elastic waves, as well as the
amplitudes of generated Rayleigh surface waves, are calculated for different
values of physical and geometrical parameters and compared with the directivity
patterns calculated in case of dipole-source representation. It is demonstrated
that the simple approach using a dipole-source representation of
laser-generated sound is rather limited, especially for description of
generated longitudinal acoustic waves. A practical criterion is established to
define the conditions under which the dipole-source representation gives
predictions with acceptable errors. It is shown that, for radiation in the
normal direction to the surface, the amplitudes of longitudinal waves are
especially sensitive to the values of thermal parameters and of the acoustic
reflection coefficient from a free solid surface. A discussion is given on the
possibility of using such a high sensitivity to the values of the reflection
coefficient for investigation of surface properties of real solids.Comment: 14 pages, 7 figure
- …
