54,040 research outputs found
Co-simulation of human digital twins and wearable inertial sensors to analyse gait event estimation
We propose a co-simulation framework comprising biomechanical human body models and wearable inertial sensor models to analyse gait events dynamically, depending on inertial sensor type, sensor positioning, and processing algorithms. A total of 960 inertial sensors were virtually attached to the lower extremities of a validated biomechanical model and shoe model. Walking of hemiparetic patients was simulated using motion capture data (kinematic simulation). Accelerations and angular velocities were synthesised according to the inertial sensor models. A comprehensive error analysis of detected gait events versus reference gait events of each simulated sensor position across all segments was performed. For gait event detection, we considered 1-, 2-, and 4-phase gait models. Results of hemiparetic patients showed superior gait event estimation performance for a sensor fusion of angular velocity and acceleration data with lower nMAEs (9%) across all sensor positions compared to error estimation with acceleration data only. Depending on algorithm choice and parameterisation, gait event detection performance increased up to 65%. Our results suggest that user personalisation of IMU placement should be pursued as a first priority for gait phase detection, while sensor position variation may be a secondary adaptation target. When comparing rotatory and translatory error components per body segment, larger interquartile ranges of rotatory errors were observed for all phase models i.e., repositioning the sensor around the body segment axis was more harmful than along the limb axis for gait phase detection. The proposed co-simulation framework is suitable for evaluating different sensor modalities, as well as gait event detection algorithms for different gait phase models. The results of our analysis open a new path for utilising biomechanical human digital twins in wearable system design and performance estimation before physical device prototypes are deployed
Cyber–Physical–Social Frameworks for Urban Big Data Systems: A Survey
The integration of things’ data on the Web and Web linking for things’ description and discovery is leading the way towards smart Cyber–Physical Systems (CPS). The data generated in CPS represents observations gathered by sensor devices about the ambient environment that can be manipulated by computational processes of the cyber world. Alongside this, the growing use of social networks offers near real-time citizen sensing capabilities as a complementary information source. The resulting Cyber–Physical–Social System (CPSS) can help to understand the real world and provide proactive services to users. The nature of CPSS data brings new requirements and challenges to different stages of data manipulation, including identification of data sources, processing and fusion of different types and scales of data. To gain an understanding of the existing methods and techniques which can be useful for a data-oriented CPSS implementation, this paper presents a survey of the existing research and commercial solutions. We define a conceptual framework for a data-oriented CPSS and detail the various solutions for building human–machine intelligence
City Data Fusion: Sensor Data Fusion in the Internet of Things
Internet of Things (IoT) has gained substantial attention recently and play a
significant role in smart city application deployments. A number of such smart
city applications depend on sensor fusion capabilities in the cloud from
diverse data sources. We introduce the concept of IoT and present in detail ten
different parameters that govern our sensor data fusion evaluation framework.
We then evaluate the current state-of-the art in sensor data fusion against our
sensor data fusion framework. Our main goal is to examine and survey different
sensor data fusion research efforts based on our evaluation framework. The
major open research issues related to sensor data fusion are also presented.Comment: Accepted to be published in International Journal of Distributed
Systems and Technologies (IJDST), 201
Closed-loop Bayesian Semantic Data Fusion for Collaborative Human-Autonomy Target Search
In search applications, autonomous unmanned vehicles must be able to
efficiently reacquire and localize mobile targets that can remain out of view
for long periods of time in large spaces. As such, all available information
sources must be actively leveraged -- including imprecise but readily available
semantic observations provided by humans. To achieve this, this work develops
and validates a novel collaborative human-machine sensing solution for dynamic
target search. Our approach uses continuous partially observable Markov
decision process (CPOMDP) planning to generate vehicle trajectories that
optimally exploit imperfect detection data from onboard sensors, as well as
semantic natural language observations that can be specifically requested from
human sensors. The key innovation is a scalable hierarchical Gaussian mixture
model formulation for efficiently solving CPOMDPs with semantic observations in
continuous dynamic state spaces. The approach is demonstrated and validated
with a real human-robot team engaged in dynamic indoor target search and
capture scenarios on a custom testbed.Comment: Final version accepted and submitted to 2018 FUSION Conference
(Cambridge, UK, July 2018
A semantic sensor web framework for proactive environmental monitoring and control.
Doctor of Philosophy in Computer Science, University of KwaZulu-Natal, Westville, 2017.Observing and monitoring of the natural and built environments is crucial for main-
taining and preserving human life. Environmental monitoring applications typically incorporate
some sensor technology to continually observe specific features of inter- est in the physical
environment and transmitting data emanating from these sensors to a computing system for analysis.
Semantic Sensor Web technology supports se- mantic enrichment of sensor data and provides
expressive analytic techniques for data fusion, situation detection and situation analysis.
Despite the promising successes of the Semantic Sensor Web technology, current Semantic
Sensor Web frameworks are typically focused at developing applications for detecting and
reacting to situations detected from current or past observations. While these reactive
applications provide a quick response to detected situations to minimize adverse effects,
they are limited when it comes to anticipating future adverse situations and determining
proactive control actions to prevent or mitigate these situations. Most current Semantic Sensor
Web frameworks lack two essential mechanisms required to achieve proactive control, namely,
mechanisms for antici- pating the future and coherent mechanisms for consistent decision
processing and planning.
Designing and developing proactive monitoring and control Semantic Sensor Web applications
is challenging. It requires incorporating and integrating different tech- niques for supporting
situation detection, situation prediction, decision making and planning in a coherent framework.
This research proposes a coherent Semantic Sen- sor Web framework for proactive monitoring and
control. It incorporates ontology
to facilitate situation detection from streaming sensor observations, statistical ma- chine
learning for situation prediction and Markov Decision Processes for decision making and
planning. The efficacy and use of the framework is evaluated through the development of two
different prototype applications. The first application is for proactive monitoring and
control of indoor air quality to avoid poor air quality situations. The second is for
proactive monitoring and control of electricity usage in blocks of residential houses to
prevent strain on the national grid. These appli- cations show the effectiveness of
the proposed framework for developing Semantic Sensor Web applications that proactively avert
unwanted environmental situations before they occur
Ambient health monitoring: the smartphone as a body sensor network component
Inertial measurement units used in commercial body sensor networks (e.g. animation suits) are inefficient, difficult to use and expensive when adapted for movement science applications concerning medical and sports science. However, due to advances in micro-electro mechanical sensors, these inertial sensors have become ubiquitous in mobile computing technologies such as smartphones. Smartphones generally use inertial sensors to enhance the interface usability. This paper investigates the use of a smartphone’s inertial sensing capability as a component in body sensor networks. It discusses several topics centered on inertial sensing: body sensor networks, smartphone networks and a prototype framework for integrating these and other heterogeneous devices. The proposed solution is a smartphone application that gathers, processes and filters sensor data for the purpose of tracking physical activity. All networking functionality is achieved by Skeletrix, a framework for gathering and organizing motion data in online repositories that are conveniently accessible to researchers, healthcare professionals and medical care workers
Context Aware Computing for The Internet of Things: A Survey
As we are moving towards the Internet of Things (IoT), the number of sensors
deployed around the world is growing at a rapid pace. Market research has shown
a significant growth of sensor deployments over the past decade and has
predicted a significant increment of the growth rate in the future. These
sensors continuously generate enormous amounts of data. However, in order to
add value to raw sensor data we need to understand it. Collection, modelling,
reasoning, and distribution of context in relation to sensor data plays
critical role in this challenge. Context-aware computing has proven to be
successful in understanding sensor data. In this paper, we survey context
awareness from an IoT perspective. We present the necessary background by
introducing the IoT paradigm and context-aware fundamentals at the beginning.
Then we provide an in-depth analysis of context life cycle. We evaluate a
subset of projects (50) which represent the majority of research and commercial
solutions proposed in the field of context-aware computing conducted over the
last decade (2001-2011) based on our own taxonomy. Finally, based on our
evaluation, we highlight the lessons to be learnt from the past and some
possible directions for future research. The survey addresses a broad range of
techniques, methods, models, functionalities, systems, applications, and
middleware solutions related to context awareness and IoT. Our goal is not only
to analyse, compare and consolidate past research work but also to appreciate
their findings and discuss their applicability towards the IoT.Comment: IEEE Communications Surveys & Tutorials Journal, 201
- …