17,418 research outputs found
Meetings and Meeting Modeling in Smart Environments
In this paper we survey our research on smart meeting rooms and its relevance for augmented reality meeting support and virtual reality generation of meetings in real time or off-line. The research reported here forms part of the European 5th and 6th framework programme projects multi-modal meeting manager (M4) and augmented multi-party interaction (AMI). Both projects aim at building a smart meeting environment that is able to collect multimodal captures of the activities and discussions in a meeting room, with the aim to use this information as input to tools that allow real-time support, browsing, retrieval and summarization of meetings. Our aim is to research (semantic) representations of what takes place during meetings in order to allow generation, e.g. in virtual reality, of meeting activities (discussions, presentations, voting, etc.). Being able to do so also allows us to look at tools that provide support during a meeting and at tools that allow those not able to be physically present during a meeting to take part in a virtual way. This may lead to situations where the differences between real meeting participants, human-controlled virtual participants and (semi-) autonomous virtual participants disappear
Mixed reality participants in smart meeting rooms and smart home enviroments
Human–computer interaction requires modeling of the user. A user profile typically contains preferences, interests, characteristics, and interaction behavior. However, in its multimodal interaction with a smart environment the user displays characteristics that show how the user, not necessarily consciously, verbally and nonverbally provides the smart environment with useful input and feedback. Especially in ambient intelligence environments we encounter situations where the environment supports interaction between the environment, smart objects (e.g., mobile robots, smart furniture) and human participants in the environment. Therefore it is useful for the profile to contain a physical representation of the user obtained by multi-modal capturing techniques. We discuss the modeling and simulation of interacting participants in a virtual meeting room, we discuss how remote meeting participants can take part in meeting activities and they have some observations on translating research results to smart home environments
Trust and Privacy Permissions for an Ambient World
Ambient intelligence (AmI) and ubiquitous computing allow us to consider a future where computation is embedded into our daily social lives. This vision raises its own important questions and augments the need to understand how people will trust such systems and at the same time achieve and maintain privacy. As a result, we have recently conducted a wide reaching study of people’s attitudes to potential AmI scenarios with a view to eliciting their privacy concerns. This chapter describes recent research related to privacy and trust with regard to ambient technology. The method used in the study is described and findings discussed
No Grice: Computers that Lie, Deceive and Conceal
In the future our daily life interactions with other people, with computers, robots and smart environments will be recorded and interpreted by computers or embedded intelligence in environments, furniture, robots, displays, and wearables. These sensors record our activities, our behavior, and our interactions. Fusion of such information and reasoning about such information makes it possible, using computational models of human behavior and activities, to provide context- and person-aware interpretations of human behavior and activities, including determination of attitudes, moods, and emotions. Sensors include cameras, microphones, eye trackers, position and proximity sensors, tactile or smell sensors, et cetera. Sensors can be embedded in an environment, but they can also move around, for example, if they are part of a mobile social robot or if they are part of devices we carry around or are embedded in our clothes or body. \ud
\ud
Our daily life behavior and daily life interactions are recorded and interpreted. How can we use such environments and how can such environments use us? Do we always want to cooperate with these environments; do these environments always want to cooperate with us? In this paper we argue that there are many reasons that users or rather human partners of these environments do want to keep information about their intentions and their emotions hidden from these smart environments. On the other hand, their artificial interaction partner may have similar reasons to not give away all information they have or to treat their human partner as an opponent rather than someone that has to be supported by smart technology.\ud
\ud
This will be elaborated in this paper. We will survey examples of human-computer interactions where there is not necessarily a goal to be explicit about intentions and feelings. In subsequent sections we will look at (1) the computer as a conversational partner, (2) the computer as a butler or diary companion, (3) the computer as a teacher or a trainer, acting in a virtual training environment (a serious game), (4) sports applications (that are not necessarily different from serious game or education environments), and games and entertainment applications
Social Intelligence Design in Ambient Intelligence
This Special Issue of AI and Society contains a selection of papers presented at the 6th Social Intelligence Design Workshop held at ITC-irst, Povo (Trento, Italy) in July 2007. Being the 6th in a series means that there now is a well-established and also a growing research area. The interest in this research area is growing because, among other things, current computing technology allows other than the traditional efficiency-oriented applications associated with computer science and interface technology. For example, in Ambient Intelligence (AmI) applications we look at sensor-equipped environments and devices (robots, smart furniture, virtual humans and pets) that support their human inhabitants during their everyday activities. These everyday activities also include computer-mediated communication, collaboration and community activities
A Role-Based Approach for Orchestrating Emergent Configurations in the Internet of Things
The Internet of Things (IoT) is envisioned as a global network of connected
things enabling ubiquitous machine-to-machine (M2M) communication. With
estimations of billions of sensors and devices to be connected in the coming
years, the IoT has been advocated as having a great potential to impact the way
we live, but also how we work. However, the connectivity aspect in itself only
accounts for the underlying M2M infrastructure. In order to properly support
engineering IoT systems and applications, it is key to orchestrate
heterogeneous 'things' in a seamless, adaptive and dynamic manner, such that
the system can exhibit a goal-directed behaviour and take appropriate actions.
Yet, this form of interaction between things needs to take a user-centric
approach and by no means elude the users' requirements. To this end,
contextualisation is an important feature of the system, allowing it to infer
user activities and prompt the user with relevant information and interactions
even in the absence of intentional commands. In this work we propose a
role-based model for emergent configurations of connected systems as a means to
model, manage, and reason about IoT systems including the user's interaction
with them. We put a special focus on integrating the user perspective in order
to guide the emergent configurations such that systems goals are aligned with
the users' intentions. We discuss related scientific and technical challenges
and provide several uses cases outlining the concept of emergent
configurations.Comment: In Proceedings of the Second International Workshop on the Internet
of Agents @AAMAS201
- …