36,446 research outputs found
Designing Awareness Support for Distributed Cooperative Design Teams
Motivation â Awareness is an integral part of remote collaborative work and has been an important theme within the CSCW research. Our project aims at understanding and mediating non-verbal cues between remote participants involved in a design project. \ud
Research approach â Within the AMIDA1 project we focus on distributed âcooperative designâ teams. We especially focus on the 'material' signals â signals in which people communicate through material artefacts, locations and their embodied actions. We apply an ethnographic approach to understand the role of physical artefacts in co-located naturalistic design setting. Based on the results we will generate important implications to support remote design work. We plan to develop a mixed-reality interface supported by a shared awareness display. This awareness display will provide information about the activities happening in the design room to remotely located participants.\ud
Findings/Design â Our preliminary investigation with real-world design teams suggests that both the materiality of designersâ work settings and their social practices play an important role in understanding these material signals that are at play. \ud
Originality/Value â Most research supporting computer mediated communication have focused on either face-to-face or linguistically oriented communication paradigms. Our research focuses on mediating the non-verbal, material cues for supporting collaborative activities without impoverishing what designers do in their day to day working lives.\ud
Take away message â An ethnographic approach allows us to understand the naturalistic practices of design teams, which can lead to designing effective technologies to support group work. In that respect, the findings of our research will have a generic value beyond the application domain chosen (design teams).\u
Collaborative Gaze Channelling for Improved Cooperation During Robotic Assisted Surgery
The use of multiple robots for performing complex tasks is becoming a common practice for many robot applications. When different operators are involved, effective cooperation with anticipated manoeuvres is important for seamless, synergistic control of all the end-effectors. In this paper, the concept of Collaborative Gaze Channelling (CGC) is presented for improved control of surgical robots for a shared task. Through eye tracking, the fixations of each operator are monitored and presented in a shared surgical workspace. CGC permits remote or physically separated collaborators to share their intention by visualising the eye gaze of their counterparts, and thus recovers, to a certain extent, the information of mutual intent that we rely upon in a vis-à -vis working setting. In this study, the efficiency of surgical manipulation with and without CGC for controlling a pair of bimanual surgical robots is evaluated by analysing the level of coordination of two independent operators. Fitts' law is used to compare the quality of movement with or without CGC. A total of 40 subjects have been recruited for this study and the results show that the proposed CGC framework exhibits significant improvement (p<0.05) on all the motion indices used for quality assessment. This study demonstrates that visual guidance is an implicit yet effective way of communication during collaborative tasks for robotic surgery. Detailed experimental validation results demonstrate the potential clinical value of the proposed CGC framework. © 2012 Biomedical Engineering Society.link_to_subscribed_fulltex
A study of event traffic during the shared manipulation of objects within a collaborative virtual environment
Event management must balance consistency and responsiveness above the requirements of shared object interaction within a Collaborative Virtual Environment
(CVE) system. An understanding of the event traffic during collaborative tasks helps in the design of all aspects of a CVE system. The application, user activity, the display
interface, and the network resources, all play a part in determining the characteristics of event management.
Linked cubic displays lend themselves well to supporting natural social human communication between remote users. To allow users to communicate naturally and subconsciously, continuous and detailed tracking is necessary. This, however, is hard to balance with the real-time consistency constraints of general shared object interaction.
This paper aims to explain these issues through a detailed examination of event traffic produced by a typical CVE, using both immersive and desktop displays, while supporting a variety of collaborative activities. We analyze event traffic during a highly collaborative task requiring various forms of shared object manipulation, including the concurrent manipulation of a shared object. Event sources are categorized and the influence of the form of object sharing as well as the display device
interface are detailed. With the presented findings the paper wishes to aid the design of future systems
Multi-Modal Human-Machine Communication for Instructing Robot Grasping Tasks
A major challenge for the realization of intelligent robots is to supply them
with cognitive abilities in order to allow ordinary users to program them
easily and intuitively. One way of such programming is teaching work tasks by
interactive demonstration. To make this effective and convenient for the user,
the machine must be capable to establish a common focus of attention and be
able to use and integrate spoken instructions, visual perceptions, and
non-verbal clues like gestural commands. We report progress in building a
hybrid architecture that combines statistical methods, neural networks, and
finite state machines into an integrated system for instructing grasping tasks
by man-machine interaction. The system combines the GRAVIS-robot for visual
attention and gestural instruction with an intelligent interface for speech
recognition and linguistic interpretation, and an modality fusion module to
allow multi-modal task-oriented man-machine communication with respect to
dextrous robot manipulation of objects.Comment: 7 pages, 8 figure
An Immersive Telepresence System using RGB-D Sensors and Head Mounted Display
We present a tele-immersive system that enables people to interact with each
other in a virtual world using body gestures in addition to verbal
communication. Beyond the obvious applications, including general online
conversations and gaming, we hypothesize that our proposed system would be
particularly beneficial to education by offering rich visual contents and
interactivity. One distinct feature is the integration of egocentric pose
recognition that allows participants to use their gestures to demonstrate and
manipulate virtual objects simultaneously. This functionality enables the
instructor to ef- fectively and efficiently explain and illustrate complex
concepts or sophisticated problems in an intuitive manner. The highly
interactive and flexible environment can capture and sustain more student
attention than the traditional classroom setting and, thus, delivers a
compelling experience to the students. Our main focus here is to investigate
possible solutions for the system design and implementation and devise
strategies for fast, efficient computation suitable for visual data processing
and network transmission. We describe the technique and experiments in details
and provide quantitative performance results, demonstrating our system can be
run comfortably and reliably for different application scenarios. Our
preliminary results are promising and demonstrate the potential for more
compelling directions in cyberlearning.Comment: IEEE International Symposium on Multimedia 201
Pointing as an Instrumental Gesture : Gaze Representation Through Indication
The research of the first author was supported by a Fulbright Visiting Scholar Fellowship and developed in 2012 during a period of research visit at the University of Memphis.Peer reviewedPublisher PD
- âŠ