4,748 research outputs found
Ubiquitous Integration and Temporal Synchronisation (UbilTS) framework : a solution for building complex multimodal data capture and interactive systems
Contemporary Data Capture and Interactive Systems (DCIS) systems are tied in with various
technical complexities such as multimodal data types, diverse hardware and software
components, time synchronisation issues and distributed deployment configurations. Building
these systems is inherently difficult and requires addressing of these complexities before the
intended and purposeful functionalities can be attained. The technical issues are often
common and similar among diverse applications.
This thesis presents the Ubiquitous Integration and Temporal Synchronisation (UbiITS)
framework, a generic solution to address the technical complexities in building DCISs. The
proposed solution is an abstract software framework that can be extended and customised to
any application requirements. UbiITS includes all fundamental software components,
techniques, system level layer abstractions and reference architecture as a collection to enable
the systematic construction of complex DCISs.
This work details four case studies to showcase the versatility and extensibility of UbiITS
framework’s functionalities and demonstrate how it was employed to successfully solve a
range of technical requirements. In each case UbiITS operated as the core element of each
application. Additionally, these case studies are novel systems by themselves in each of their
domains. Longstanding technical issues such as flexibly integrating and interoperating
multimodal tools, precise time synchronisation, etc., were resolved in each application by
employing UbiITS. The framework enabled establishing a functional system infrastructure in
these cases, essentially opening up new lines of research in each discipline where these
research approaches would not have been possible without the infrastructure provided by the
framework. The thesis further presents a sample implementation of the framework on a
device firmware exhibiting its capability to be directly implemented on a hardware platform.
Summary metrics are also produced to establish the complexity, reusability, extendibility,
implementation and maintainability characteristics of the framework.Engineering and Physical Sciences Research Council (EPSRC) grants - EP/F02553X/1, 114433 and 11394
Multimodal fusion : gesture and speech input in augmented reality environment
Augmented Reality (AR) has the capability to interact with the virtual objects and physical objects simultaneously since it combines the real world with virtual world seamlessly. However, most AR interface applies conventional Virtual Reality (VR) interaction techniques without modification. In this paper we explore the multimodal fusion for AR with speech and hand gesture input. Multimodal fusion enables users to interact with computers through various input modalities like speech, gesture, and eye gaze. At the first stage to propose the multimodal interaction, the input modalities are decided to be selected before be integrated in an interface. The paper presents several related works about to recap the multimodal approaches until it recently has been one of the research trends in AR. It presents the assorted existing works in multimodal for VR and AR. In AR, multimodal considers as the solution to improve the interaction between the virtual and physical entities. It is an ideal interaction technique for AR applications since AR supports interactions in real and virtual worlds in the real-time. This paper describes the recent studies in AR developments that appeal gesture and speech inputs. It looks into multimodal fusion and its developments, followed by the conclusion.This paper will give a guideline on multimodal fusion on how to integrate the gesture and speech inputs in AR environment
- …