46 research outputs found
Shifting attention in viewer- and object-based reference frames after unilateral brain injury
The aims of the present study were to investigate the respective roles that object- and viewer-based reference frames play in reorienting visual attention, and to assess their influence after unilateral brain injury. To do so, we studied 16 right hemisphere injured (RHI) and 13 left hemisphere injured (LHI) patients. We used a cueing design that manipulates the location of cues and targets relative to a display comprised of two rectangles (i.e., objects). Unlike previous studies with patients, we presented all cues at midline rather than in the left or right visual fields. Thus, in the critical conditions in which targets were presented laterally, reorienting of attention was always from a midline cue. Performance was measured for lateralized target detection as a function of viewer-based (contra- and ipsilesional sides) and object-based (requiring reorienting within or between objects) reference frames. As expected, contralesional detection was slower than ipsilesional detection for the patients. More importantly, objects influenced target detection differently in the contralesional and ipsilesional fields. Contralesionally, reorienting to a target within the cued object took longer than reorienting to a target in the same location but in the uncued object. This finding is consistent with object-based neglect. Ipsilesionally, the means were in the opposite direction. Furthermore, no significant difference was found in object-based influences between the patient groups (RHI vs. LHI). These findings are discussed in the context of reference frames used in reorienting attention for target detection
Recommended from our members
Mismatch Negativity (MMN) reveals inefficient auditory ventral stream function in chronic auditory comprehension impairments
Background: Auditory discrimination is significantly impaired in Wernicke’s aphasia (WA) and thought to be causatively related to the language comprehension impairment which characterises the condition. This study used mismatch negativity (MMN) to investigate the neural responses corresponding to successful and impaired auditory discrimination in WA.
Methods: Behavioural auditory discrimination thresholds of CVC syllables and pure tones were measured in WA (n=7) and control (n=7) participants. Threshold results were used to develop multiple-deviant mismatch negativity (MMN) oddball paradigms containing deviants which were either perceptibly or non-perceptibly different from the standard stimuli. MMN analysis investigated differences associated with group, condition and perceptibility as well as the relationship between MMN responses and comprehension (within which behavioural auditory discrimination profiles were examined).
Results: MMN waveforms were observable to both perceptible and non-perceptible auditory changes. Perceptibility was only distinguished by MMN amplitude in the PT condition. The WA group could be distinguished from controls by an increase in MMN response latency to CVC stimuli change. Correlation analyses displayed relationship between behavioural CVC discrimination and MMN amplitude in the control group, where greater amplitude corresponded to better discrimination. The WA group displayed the inverse effect; both discrimination accuracy and auditory comprehension scores were reduced with increased MMN amplitude. In the WA group, a further correlation was observed between the lateralisation of MMN response and CVC discrimination accuracy; the greater the bilateral involvement the better the discrimination accuracy.
Conclusions: The results from this study provide further evidence for the nature of auditory comprehension impairment in WA and indicate that the auditory discrimination deficit is grounded in a reduced ability to engage in efficient hierarchical processing and the construction of invariant auditory objects. Correlation results suggest that people with chronic WA may rely on an inefficient, noisy right hemisphere auditory stream when attempting to process speech stimuli
Disentangling neural processing of masked and masking stimulus by means of event-related contralateral – ipsilateral differences of EEG potentials
In spite of the excellent temporal resolution of event-related EEG potentials
(ERPs), the overlapping potentials evoked by masked and masking stimuli are hard
to disentangle. However, when both masked and masking stimuli consist of pairs
of relevant and irrelevant stimuli, one left and one right from fixation, with
the side of the relevant element varying between pairs, effects of masked and
masking stimuli can be distinguished by means of the contralateral preponderance
of the potentials evoked by the relevant elements, because the relevant elements
may independently change sides in masked and masking stimuli. Based on a
reanalysis of data from which only selected contralateral-ipsilateral effects
had been previously published, the present contribution will provide a more
complete picture of the ERP effects in a masked-priming task. Indeed, effects
evoked by masked primes and masking targets heavily overlapped in conventional
ERPs and could be disentangled to a certain degree by contralateral-ipsilateral
differences. Their major component, the N2pc, is interpreted as indicating
preferential processing of stimuli matching the target template, which process
can neither be identified with conscious perception nor with shifts of spatial
attention. The measurements showed that the triggering of response preparation
by the masked stimuli did not depend on their discriminability, and their
priming effects on the processing of the following target stimuli were
qualitatively different for stimulus identification and for response
preparation. These results provide another piece of evidence for the
independence of motor-related and perception-related effects of masked
stimuli
Fast Detection of Unexpected Sound Intensity Decrements as Revealed by Human Evoked Potentials
The detection of deviant sounds is a crucial function of the auditory system and is reflected by the automatically elicited mismatch negativity (MMN), an auditory evoked potential at 100 to 250 ms from stimulus onset. It has recently been shown that rarely occurring frequency and location deviants in an oddball paradigm trigger a more negative response than standard sounds at very early latencies in the middle latency response of the human auditory evoked potential. This fast and early ability of the auditory system is corroborated by the finding of neurons in the animal auditory cortex and subcortical structures, which restore their adapted responsiveness to standard sounds, when a rare change in a sound feature occurs. In this study, we investigated whether the detection of intensity deviants is also reflected at shorter latencies than those of the MMN. Auditory evoked potentials in response to click sounds were analyzed regarding the auditory brain stem response, the middle latency response (MLR) and the MMN. Rare stimuli with a lower intensity level than standard stimuli elicited (in addition to an MMN) a more negative potential in the MLR at the transition from the Na to the Pa component at circa 24 ms from stimulus onset. This finding, together with the studies about frequency and location changes, suggests that the early automatic detection of deviant sounds in an oddball paradigm is a general property of the auditory system
Spatio-Temporal Features of Visual Exploration in Unilaterally Brain-Damaged Subjects with or without Neglect: Results from a Touchscreen Test
Cognitive assessment in a clinical setting is generally made by pencil-and-paper tests, while computer-based tests enable the measurement and the extraction of additional performance indexes. Previous studies have demonstrated that in a research context exploration deficits occur also in patients without evidence of unilateral neglect at pencil-and-paper tests. The objective of this study is to apply a touchscreen-based cancellation test, feasible also in a clinical context, to large groups of control subjects and unilaterally brain-damaged patients, with and without unilateral spatial neglect (USN), in order to assess disturbances of the exploratory skills. A computerized cancellation test on a touchscreen interface was used for assessing the performance of 119 neurologically unimpaired control subjects and 193 patients with unilateral right or left hemispheric brain damage, either with or without USN. A set of performance indexes were defined including Latency, Proximity, Crossings and their spatial lateral gradients, and Preferred Search Direction. Classic outcome scores were computed as well. Results show statistically significant differences among groups (assumed p<0.05). Right-brain-damaged patients with USN were significantly slower (median latency per detected item was 1.18 s) and less efficient (about 13 search-path crossings) in the search than controls (median latency 0.64 s; about 3 crossings). Their preferred search direction (53.6% downward, 36.7% leftward) was different from the one in control patients (88.2% downward, 2.1% leftward). Right-brain-damaged patients without USN showed a significantly abnormal behavior (median latency 0.84 s, about 5 crossings, 83.3% downward and 9.1% leftward direction) situated half way between controls and right-brain-damaged patients with USN. Left-brain-damaged patients without USN were significantly slower and less efficient than controls (latency 1.19 s, about 7 crossings), preserving a normal preferred search direction (93.7% downward). Therefore, the proposed touchscreen-based assessment had evidenced disorders in spatial exploration also in patients without clinically diagnosed USN
Auditory event-related potentials
Auditory event related potentials are electric potentials (AERP, AEP) and magnetic fields (AEF) generated by the synchronous activity of large neural populations in the brain, which are time-locked to some actual or expected sound event
Mechanisms of Implicit Timing: Cognitive and Electrophysiological Manifestations of Temporal Expectations Following Rhythmic Visual Input
Temporal processing is frequently performed implicitly, when temporal calculations are required in order to perform a task, whose output is not temporal in essence. One prevalent use of implicit timing is in the formation of temporal expectations, in which the timing of an upcoming event is predicted in order to prepare for its occurrence. The sources of this prediction can be “endogenous”, relying on memorizing the temporal interval between the expected event and a preceding reference event, measuring the time from the onset of the reference, and increasing preparedness when the expected interval elapses. However, temporal expectations can also be “exogenous”, namely based on extracting temporal information from the ongoing temporal dynamics of the input, rather than memory-dependent. One unique case of such temporal dynamics is when the input is isochronous (i.e., stimuli appear rhythmically, with fixed inter-onset-interval), such as in biological motion, speech, or music. Based on findings from auditory rhythms, it has been suggested that in such case, expectations are realized by synchronization of internal oscillators to the input periodicity, a process that encompasses both the representation of the interval and the application of expectation at the correct time. In two studies, we investigate temporal expectations created in visual rhythmic context, showing that they have unique cognitive and electrophysiological characteristics, and dissociating them from other types of temporal expectations
When Synchronizing to Rhythms Is Not a Good Thing: Modulations of Preparatory and Post-Target Neural Activity When Shifting Attention Away from On-Beat Times of a Distracting Rhythm
Environmental rhythms potently drive predictive resource allocation in time, typically leading to perceptual and motor benefits for on-beat, relative to off-beat, times, even if the rhythmic stream is not intentionally used. In two human EEG experiments, we investigated the behavioral and electrophysiological expressions of using rhythms to direct resources away from on-beat times. This allowed us to distinguish goal-directed attention from the automatic capture of attention by rhythms. The following three conditions were compared: (1) a rhythmic stream with targets appearing frequently at a fixed off-beat position; (2) a rhythmic stream with targets appearing frequently at on-beat times; and (3) a nonrhythmic stream with matched target intervals. Shifting resources away from on-beat times was expressed in the slowing of responses to on-beat targets, but not in the facilitation of off-beat targets. The shifting of resources was accompanied by anticipatory adjustment of the contingent negative variation (CNV) buildup toward the expected off-beat time. In the second experiment, off-beat times were jittered, resulting in a similar CNV adjustment and also in preparatory amplitude reduction of beta-band activity. Thus, the CNV and beta activity track the relevance of time points and not the rhythm, given sufficient incentive. Furthermore, the effects of task relevance (appearing in a task-relevant vs irrelevant time) and rhythm (appearing on beat vs off beat) had additive behavioral effects and also dissociable neural manifestations in target-evoked activity: rhythm affected the target response as early as the P1 component, while relevance affected only the later N2 and P3. Thus, these two factors operate by distinct mechanisms
Automatic bias of temporal expectations following temporally regular input independently of high-level temporal expectation
Exposure to rhythmic stimulation results in facilitated responses to events that appear in-phase with the rhythm and modulation of anticipatory and target-evoked brain activity, presumably reflecting "exogenous," unintentional temporal expectations. However, the extent to which this effect is independent from intentional processes is not clear. In two EEG experiments, we isolated the unintentional component of this effect from high-level, intentional factors. Visual targets were presented either in-phase or out-of-phase with regularly flickering colored stimuli. In different blocks, the rhythm could be predictive (i.e., high probability for in-phase target) or not, and the color could be predictive (i.e., validly cue the interval to the target) or not. Exposure to nonpredictive rhythms resulted in faster responses for in-phase targets, even when the color predicted specific out-of-phase target times. Also, the contingent negative variation, an EEG component reflecting temporal anticipation, followed the interval of the nonpredictive rhythm and not that of the predictive color. Thus, rhythmic stimulation unintentionally induced expectations, even when this was detrimental. Intentional usage of predictive rhythms to form expectations resulted in a stronger behavioral effect, and only predictive cues modulated the latency of the target-evoked P3, presumably reflecting stimulus evaluation. These findings establish the existence of unintentional temporal expectations in rhythmic contexts, dissociate them from intentional expectations, and highlight the need to distinguish between the source of expectation (exogenous-endogenous) and the level of voluntary control involved in it (unintentional-intentional)
Integration without awareness: expanding the limits of unconscious processing
Human conscious awareness is commonly seen as the climax of evolution. However, what function-if any-it serves in human behavior is still debated. One of the leading suggestions is that the cardinal function of conscious awareness is to integrate numerous inputs-including the multitude of features and objects in a complex scene-across different levels of analysis into a unified, coherent, and meaningful perceptual experience. Here we demonstrate, however, that integration of objects with their background scenes can be achieved without awareness of either. We used a binocular rivalry technique known as continuous flash suppression to induce perceptual suppression in a group of human observers. Complex scenes that included incongruent objects escaped perceptual suppression faster than normal scenes did. We conclude that visual awareness is not needed for object-background integration or for processing the likelihood of an object to appear within a given semantic context, but may be needed for dealing with novel situations