1,251 research outputs found
uC: Ubiquitous Collaboration Platform for Multimodal Team Interaction Support
A human-centered computing platform that improves teamwork and transforms the “human- computer interaction experience” for distributed teams is presented. This Ubiquitous Collaboration, or uC (“you see”), platform\u27s objective is to transform distributed teamwork (i.e., work occurring when teams of workers and learners are geographically dispersed and often interacting at different times). It achieves this goal through a multimodal team interaction interface realized through a reconfigurable open architecture. The approach taken is to integrate: (1) an intuitive speech- and video-centric multi-modal interface to augment more conventional methods (e.g., mouse, stylus and touch), (2) an open and reconfigurable architecture supporting information gathering, and (3) a machine intelligent approach to analysis and management of heterogeneous live and stored sensor data to support collaboration. The system will transform how teams of people interact with computers by drawing on both the virtual and physical environment
Interim research assessment 2003-2005 - Computer Science
This report primarily serves as a source of information for the 2007 Interim Research Assessment Committee for Computer Science at the three technical universities in the Netherlands. The report also provides information for others interested in our research activities
Developing a Mobile Learning Environment: An Axiomatic Approach
A new mobile environment for learning has been designed via an axiomatic approach. And by simultaneously designing both tools (software) and processes (pedagogy), the resulting environment matches the functional requirements of the instructional program. This paper describes the axioms established for mobile learning as well as development of the mobile computing environment. The paper discusses the developmental evolution and system architecture as well as the requirements of the portable training programs being offered via this new system. Apps are designed to connect learners, instructors and practitioners as well as to facilitate collaborative learning from a variety of mobile devices, anywhere in the World. Future content, apps and systems development will connect the physical and virtual environments, in order to truly enhance the mobile learning experience for people on the move
Immersive Insights: A Hybrid Analytics System for Collaborative Exploratory Data Analysis
In the past few years, augmented reality (AR) and virtual reality (VR)
technologies have experienced terrific improvements in both accessibility and
hardware capabilities, encouraging the application of these devices across
various domains. While researchers have demonstrated the possible advantages of
AR and VR for certain data science tasks, it is still unclear how these
technologies would perform in the context of exploratory data analysis (EDA) at
large. In particular, we believe it is important to better understand which
level of immersion EDA would concretely benefit from, and to quantify the
contribution of AR and VR with respect to standard analysis workflows.
In this work, we leverage a Dataspace reconfigurable hybrid reality
environment to study how data scientists might perform EDA in a co-located,
collaborative context. Specifically, we propose the design and implementation
of Immersive Insights, a hybrid analytics system combining high-resolution
displays, table projections, and augmented reality (AR) visualizations of the
data.
We conducted a two-part user study with twelve data scientists, in which we
evaluated how different levels of data immersion affect the EDA process and
compared the performance of Immersive Insights with a state-of-the-art,
non-immersive data analysis system.Comment: VRST 201
Semantic-aware Digital Twin for Metaverse: A Comprehensive Review
To facilitate the deployment of digital twins in Metaverse, the paradigm with
semantic awareness has been proposed as a means for enabling accurate and
task-oriented information extraction with inherent intelligence. However, this
framework requires all devices in the Metaverse environment to be directly
linked with the semantic model to enable faithful interpretation of messages.
In contrast, this article introduces the digital twin framework, considering a
smart industrial application, which enables semantic communication in
conjugation with the Metaverse enabling technologies. The fundamentals of this
framework are demonstrated on an industrial shopfloor management use case with
a digital twin so as to improve its performance through semantic communication.
An overview of semantic communication, Metaverse, and digital twins is
presented. Integration of these technologies with the basic architecture as
well as the impact on future industrial applications is presented. In a
nutshell, this article showcases how semantic awareness can be an effective
candidate in the implementation of digital twins for Metaverse applications.Comment: 9 pages, 5 figures, 1 tabl
Emotion in Future Intelligent Machines
Over the past decades, research in cognitive and affective neuroscience has
emphasized that emotion is crucial for human intelligence and in fact
inseparable from cognition. Concurrently, there has been a significantly
growing interest in simulating and modeling emotion in robots and artificial
agents. Yet, existing models of emotion and their integration in cognitive
architectures remain quite limited and frequently disconnected from
neuroscientific evidence. We argue that a stronger integration of emotion in
robot models is critical for the design of intelligent machines capable of
tackling real world problems. Drawing from current neuroscientific knowledge,
we provide a set of guidelines for future research in artificial emotion and
intelligent machines more generally
Recommended from our members
Designing an AR interface to improve trust in Human-Robots collaboration
In a global, e-commerce marketplace, product customisation is driven towards manufacturing flexibility. Conventional caged robots are designed for high volume and low mix production cannot always comply with the increasing low volume and high customisation requirements. In this scenario, the interest in collaborative robots is growing. A critical aspect of Human-Robot Collaboration (HRC) is human trust in robots. This research focuses on increasing the human confidence and trust in robots by designing an Augmented Reality (AR) interface for HRC. The variable affecting the trust involved in HRC have been estimated. These have been utilised for designing the AR-HRC. The proposed design aims to provide situational awareness and spatial dialog. The AR-HRC developed has been tested on 15 participants which have performed a “pick-and-place” task. The results show that the utilisation of AR in the proposed scenario positively affects the human trust in robot. The human-robot collaboration enhanced by AR are more natural and effective. The trust has been measured through an empirical psychometric method also presented in this paper
- …