55 research outputs found

    Collaborating in spatial tasks: Partners adapt the perspective of their descriptions, coordination strategies, and memory representations.

    Get PDF
    The partner’s viewpoint influences spatial descriptions and, when strongly emphasized, spatial memories as well. We examined whether partnerspecific information affects the representations people spontaneously construct, the description strategies they spontaneously select, and the representations their collaborating partner constructs based on these descriptions. Directors described to a misaligned Matcher arrays learned while either knowing the Matcher’s viewpoint or not. Knowing the Matcher’s viewpoint led to distinctive processing in spatial judgments and a rotational bias in array drawings. Directors’ descriptions reflected strategic choices, suggesting that partners considered each other’s computational demands. Such strategies were effective as reflected by the number of conversational turns partners took to coordinate. Matchers represented both partners’ viewpoints in memory, with the Directors’ descriptions predicting the facilitated perspective. Thus, partners behave contingently in spatial tasks to optimize their coordination: the availability of the partner’s viewpoint influences one’s memory and description strategies, which in turn influence the partner’s memory

    Social and representational cues jointly influence spatial perspective-taking.

    Get PDF
    We examined how social cues (the conversational partner’s viewpoint) and representational ones (the intrinsic structure of a spatial layout) jointly shape people’s spatial memory representations and their subsequent descriptions. In 24 pairs, Directors studied an array with a symmetrical structure while either knowing their Matcher’s subsequent viewpoint or not. During the subsequent description of the array, the array’s intrinsic structure was aligned with the Director, the Matcher, or neither partner. According to memory tests preceding descriptions, Directors who had studied the array while aligned with its structure were more likely to use its orientation as an organizing direction. Directors who had studied the array while misaligned with its structure used its orientation more frequently as an organizing orientation when knowing that the Matcher would be aligned with it, but used their own viewpoint more frequently as an organizing direction when not knowing the Matcher’s viewpoint. Directors also adapted their descriptions strategically, using more egocentric expressions when aligned with the intrinsic structure and more partner-centered expressions when their Matchers were the ones aligned with the structure, even when this information wasn’t available in advance. These findings suggest that speakers are guided by converging social and representational cues to adapt flexibly the organization of their memories and the perspectives of their descriptions

    What's so difficult with adopting imagined perspectives?

    Get PDF
    Research on spatial cognition suggests that transformation processes and/or spatial conflicts may influence performance on mental perspective-taking tasks. However, conflicting findings have complicated our understanding about the processes involved in perspectivetaking, particularly those giving rise to angular disparity effects, whereby performance worsens as the imagined perspective adopted deviates from one’s actual perspective. Based on data from experiments involving mental perspective- taking in immediate and remote spatial layouts, we propose here a novel account for the difficulty with perspective-taking. According to this account, the main difficulty lies in maintaining an imagined perspective in working memory, especially in the presence of salient sensorimotor information

    The protagonist's first perspective influences the encoding of spatial information in narratives

    Get PDF
    Three experiments examined the first-perspective alignment effect that is observed when retrieving spatial information from memory about described environments. Participants read narratives that described the viewpoint of a protagonist in fictitious environments and then pointed to the memorized locations of described objects from imagined perspectives. Results from Experiments 1 and 2 showed that performance was best when participants responded from the protagonist’s first perspective even though object locations were described from a different perspective. In Experiment 3, in which participants were physically oriented with the perspective used to describe object locations, performance from that description perspective was better than that from the protagonist’s first perspective, which was, in turn, better than performance from other perspectives. These findings suggest that when reading narratives, people default to using a reference frame that is aligned with their own facing direction, although physical movement may facilitate retrieval from other perspectives

    Integration of spatial relations across perceptual experiences.

    Get PDF
    People often carry out tasks that entail coordinating spatial information encoded in temporally and/or spatially distinct perceptual experiences. Much research has been conducted to determine whether such spatial information is integrated into a single spatial representation or whether it is kept in separate representations that can be related at the time of retrieval. Here, we review the existing literature on the integration of spatial information and present results from a new experiment aimed at examining whether locations encoded from different perspectives in the same physical environments are integrated into a single spatial representation. Overall, our findings, coupled with those from other studies, suggest that separate spatial representations are maintained in memory

    When gestures show us the way: Co-speech gestures selectively facilitate navigation and spatial memory.

    Get PDF
    How does gesturing during route learning relate to subsequent spatial performance? We examined the relationship between gestures produced spontaneously while studying route directions and spatial representations of the navigated environment. Participants studied route directions, then navigated those routes from memory in a virtual environment, and finally had their memory of the environment assessed. We found that, for navigators with low spatial perspective-taking performance on the Spatial Orientation Test, more gesturing from a survey perspective predicted more accurate memory following navigation. Thus, co-thought gestures accompanying route learning relate to performance selectively, depending on the gesturers’ spatial ability and the perspective of their gestures. Survey gestures may help some individuals visualize an overall route that they can retain in memory

    The conversational partner's perspective affects spatial memory and descriptions.

    Get PDF
    We examined whether people spontaneously represent the partner’s viewpoint in spatial memory when it is available in advance and whether they adapt their spontaneous descriptions accordingly. In 18 pairs, Directors studied arrays of objects while: (1) not knowing about having to describe the array to a Matcher, (2) knowing about the subsequent description, and (3) knowing the Matcher’s subsequent viewpoint, which was offset by 90, 135, or 180. In memory tests preceding descriptions, Directors represented the Matcher’s viewpoint when it was known during study, taking longer to imagine orienting to perspectives aligned with it and rotating their drawings of arrays toward it. Conversely, when Directors didn’t know their Matcher’s viewpoint, they encoded arrays egocentrically, being faster to imagine orienting to and to respond from perspectives aligned with their own. Directors adapted their descriptions flexibly, using partner-centered spatial expressions more frequently when misaligned by 90 and egocentric ones when misaligned by 135. Knowing their misalignment in advance helped partners recognize when descriptions would be most difficult for Directors (at 135) and to mutually agree on using their perspective. Thus, in collaborative tasks, people don’t rely exclusively on their spatial memory but also use other pertinent perceptual information (e.g., their misalignment from their partner) to assess the computational demands on each partner and select strategies that maximize the efficiency of communication

    Virtual Reality and Symptoms Management of Anxiety, Depression, Fatigue, and Pain: A Systematic Review

    Get PDF
    In recent years, virtual reality (VR) has become an interesting alternative to traditional exposure-based therapies for many symptoms. VR involves immersion in a computer-generated virtual environment that minimizes avoidance and facilitates emotional processing. The objective of this systematic review is to evaluate evidence on the intervention effect of VR on anxiety, depression, fatigue, and pain. The research strategy of this systematic review included three electronic databases (MEDLINE/PubMed, Cochrane Library, and ScienceDirect) based on predetermined inclusion and exclusion criteria. Published quantitative studies from 2000 to 2020 were identified, which examined the effect of VR intervention on four different symptoms related to symptoms experienced by cancer patients. Quality assessments, data extractions, and analysis were completed on all included studies. A total of 882 titles and abstracts were screened, and 23 studies were included in the review. The studies were grouped according to the symptoms: anxiety and depression, fatigue, and pain. The review showed that VR intervention is more effective compared with the control (i.e., standard care) for anxiety, depression, fatigue, and pain. VR can reduce effectively these symptoms in different contexts and diseases, including cancer. The evidence suggests that there is value in exploring this intervention as a potential crossover treatment for these symptoms in patients. This study contributes to evidence that distraction is an effective symptom management mechanism. The findings are congruent with the theoretical framework, supporting the premise that VR, as an emotion-focused distraction intervention, decreases the severity of these symptoms

    Sensorimotor alignment effects in the learning environment and in novel environments.

    Full text link
    Four experiments investigated the conditions contributing to sensorimotor alignment effects (i.e., the advantage for spatial judgments from imagined perspectives aligned with the body). Through virtual reality technology, participants learned object locations around a room (learning room) and made spatial judgments from imagined perspectives aligned or misaligned with their actual facing direction. Senso-rimotor alignment effects were found when testing occurred in the learning room but not after walking 3 m into a neighboring (novel) room. Sensorimotor alignment effects returned after returning to the learning room or after providing participants with egocentric imagery instructions in the novel room. Additionally, visual and spatial similarities between the test and learning environments were indepen-dently sufficient to cause sensorimotor alignment effects. Memory alignment effects, independent from sensorimotor alignment effects, occurred in all testing conditions. Results are interpreted in the context of two-system spatial memory theories positing separate representations to account for sensorimotor and memory alignment effects
    • …
    corecore