1,121 research outputs found
The phonetics of second language learning and bilingualism
This chapter provides an overview of major theories and findings in the field of second language (L2) phonetics and phonology. Four main conceptual frameworks are discussed and compared: the Perceptual Assimilation Model-L2, the Native Language Magnet Theory, the Automatic Selection Perception Model, and the Speech Learning Model. These frameworks differ in terms of their empirical focus, including the type of learner (e.g., beginner vs. advanced) and target modality (e.g., perception vs. production), and in terms of their theoretical assumptions, such as the basic unit or window of analysis that is relevant (e.g., articulatory gestures, position-specific allophones). Despite the divergences among these theories, three recurring themes emerge from the literature reviewed. First, the learning of a target L2 structure (segment, prosodic pattern, etc.) is influenced by phonetic and/or phonological similarity to structures in the native language (L1). In particular, L1-L2 similarity exists at multiple levels and does not necessarily benefit L2 outcomes. Second, the role played by certain factors, such as acoustic phonetic similarity between close L1 and L2 sounds, changes over the course of learning, such that advanced learners may differ from novice learners with respect to the effect of a specific variable on observed L2 behavior. Third, the connection between L2 perception and production (insofar as the two are hypothesized to be linked) differs significantly from the perception-production links observed in L1 acquisition. In service of elucidating the predictive differences among these theories, this contribution discusses studies that have investigated L2 perception and/or production primarily at a segmental level. In addition to summarizing the areas in which there is broad consensus, the chapter points out a number of questions which remain a source of debate in the field today.https://drive.google.com/open?id=1uHX9K99Bl31vMZNRWL-YmU7O2p1tG2wHhttps://drive.google.com/open?id=1uHX9K99Bl31vMZNRWL-YmU7O2p1tG2wHhttps://drive.google.com/open?id=1uHX9K99Bl31vMZNRWL-YmU7O2p1tG2wHAccepted manuscriptAccepted manuscrip
How learning to abstract shapes neural sound representations
The transformation of acoustic signals into abstract perceptual representations is the essence of the efficient and goal-directed neural processing of sounds in complex natural environments. While the human and animal auditory system is perfectly equipped to process the spectrotemporal sound features, adequate sound identification and categorization require neural sound representations that are invariant to irrelevant stimulus parameters. Crucially, what is relevant and irrelevant is not necessarily intrinsic to the physical stimulus structure but needs to be learned over time, often through integration of information from other senses. This review discusses the main principles underlying categorical sound perception with a special focus on the role of learning and neural plasticity. We examine the role of different neural structures along the auditory processing pathway in the formation of abstract sound representations with respect to hierarchical as well as dynamic and distributed processing models. Whereas most fMRI studies on categorical sound processing employed speech sounds, the emphasis of the current review lies on the contribution of empirical studies using natural or artificial sounds that enable separating acoustic and perceptual processing levels and avoid interference with existing category representations. Finally, we discuss the opportunities of modern analyses techniques such as multivariate pattern analysis (MVPA) in studying categorical sound representations. With their increased sensitivity to distributed activation changesâeven in absence of changes in overall signal levelâthese analyses techniques provide a promising tool to reveal the neural underpinnings of perceptually invariant sound representations
Building phonetic categories: an argument for the role of sleep
The current review provides specific predictions for the role of sleep-mediated memory consolidation in the formation of new speech sound representations. Specifically, this discussion will highlight selected literature on the different ideas concerning category representation in speech, followed by a broad overview of memory consolidation and how it relates to human behavior, as relevant to speech/perceptual learning. In combining behavioral and physiological accounts from animal models with insights from the human consolidation literature on auditory skill/word learning, we are in the early stages of understanding how the transfer of experiential information between brain structures during sleep manifests in changes to online perception. Arriving at the conclusion that this process is crucial in perceptual learning and the formation of novel categories, further speculation yields the adjacent claim that the habitual disruption in this process leads to impoverished quality in the representation of speech sounds
Language-specificity in auditory perception of Chinese tones
PL1213, LoC Subject Headings: Auditory perception, Chinese language--Tone, Chinese language--Phonolog
Do Infants Really Learn Phonetic Categories?
Early changes in infantsâ ability to perceive native and nonnative speech sound contrasts are typically attributed to their developing knowledge of phonetic categories. We critically examine this hypothesis and argue that there is little direct evidence of category knowledge in infancy. We then propose an alternative account in which infantsâ perception changes because they are learning a perceptual space that is appropriate to represent speech, without yet carving up that space into phonetic categories. If correct, this new account has substantial implications for understanding early language development
Idiosyncratic use of bottom-up and top-down information leads to differences in speech perception flexibility: Converging evidence from ERPs and eye-tracking
Available online 8 October 2021.Listeners generally categorize speech sounds in a gradient manner. However, recent work, using a visual
analogue scaling (VAS) task, suggests that some listeners show more categorical performance, leading to less
flexible cue integration and poorer recovery from misperceptions (Kapnoula et al., 2017, 2021). We asked how
individual differences in speech gradiency can be reconciled with the well-established gradiency in the modal
listener, showing how VAS performance relates to both Visual World Paradigm and EEG measures of gradiency.
We also investigated three potential sources of these individual differences: inhibitory control; lexical inhibition;
and early cue encoding. We used the N1 ERP component to track pre-categorical encoding of Voice Onset Time
(VOT). The N1 linearly tracked VOT, reflecting a fundamentally gradient speech perception; however, for less
gradient listeners, this linearity was disrupted near the boundary. Thus, while all listeners are gradient, they may
show idiosyncratic encoding of specific cues, affecting downstream processing.This project was supported by NIH Grant
DC008089 awarded to BM. This project has received funding from the
European Unionâs Horizon 2020 research and innovation programme
under the Marie SkĆodowska-Curie grant agreement No 793919, awarded
to EK. This work was partially supported by the Basque Government
through the BERC 2018-2021 program and by the Spanish State
Research Agency through BCBL Severo Ochoa excellence accreditation
SEV-2015-0490
- âŠ