2,284 research outputs found
Towards a Practical Pedestrian Distraction Detection Framework using Wearables
Pedestrian safety continues to be a significant concern in urban communities
and pedestrian distraction is emerging as one of the main causes of grave and
fatal accidents involving pedestrians. The advent of sophisticated mobile and
wearable devices, equipped with high-precision on-board sensors capable of
measuring fine-grained user movements and context, provides a tremendous
opportunity for designing effective pedestrian safety systems and applications.
Accurate and efficient recognition of pedestrian distractions in real-time
given the memory, computation and communication limitations of these devices,
however, remains the key technical challenge in the design of such systems.
Earlier research efforts in pedestrian distraction detection using data
available from mobile and wearable devices have primarily focused only on
achieving high detection accuracy, resulting in designs that are either
resource intensive and unsuitable for implementation on mainstream mobile
devices, or computationally slow and not useful for real-time pedestrian safety
applications, or require specialized hardware and less likely to be adopted by
most users. In the quest for a pedestrian safety system that achieves a
favorable balance between computational efficiency, detection accuracy, and
energy consumption, this paper makes the following main contributions: (i)
design of a novel complex activity recognition framework which employs motion
data available from users' mobile and wearable devices and a lightweight
frequency matching approach to accurately and efficiently recognize complex
distraction related activities, and (ii) a comprehensive comparative evaluation
of the proposed framework with well-known complex activity recognition
techniques in the literature with the help of data collected from human subject
pedestrians and prototype implementations on commercially-available mobile and
wearable devices
Recommended from our members
Context-awareness for mobile sensing: a survey and future directions
The evolution of smartphones together with increasing computational power have empowered developers to create innovative context-aware applications for recognizing user related social and cognitive activities in any situation and at any location. The existence and awareness of the context provides the capability of being conscious of physical environments or situations around mobile device users. This allows network services to respond proactively and intelligently based on such awareness. The key idea behind context-aware applications is to encourage users to collect, analyze and share local sensory knowledge in the purpose for a large scale community use by creating a smart network. The desired network is capable of making autonomous logical decisions to actuate environmental objects, and also assist individuals. However, many open challenges remain, which are mostly arisen due to the middleware services provided in mobile devices have limited resources in terms of power, memory and bandwidth. Thus, it becomes critically important to study how the drawbacks can be elaborated and resolved, and at the same time better understand the opportunities for the research community to contribute to the context-awareness. To this end, this paper surveys the literature over the period of 1991-2014 from the emerging concepts to applications of context-awareness in mobile platforms by providing up-to-date research and future research directions. Moreover, it points out the challenges faced in this regard and enlighten them by proposing possible solutions
An efficient human activity recognition model based on deep learning approaches
Human Activity Recognition (HAR) has gained traction in recent years in diverse areas such as observation, entertainment, teaching and healthcare, using wearable and smartphone sensors. Such environments and systems necessitate and subsume activity recognition, aimed at recognizing the actions, characteristics, and goals of one or more individuals from a temporal series of observations streamed from one or more sensors. Different developed models for HAR have been explained in literature. Deep learning systems and algorithms were shown to perform highly in HAR in recent years, but these algorithms need lots of computerization to be deployed efficiently in applications. This paper presents a HAR lightweight, low computing capacity, deep learning model, which is ideal for use in real-time applications. The generic HAR framework for smartphone sensor data is proposed, based on Long Short-Term Memory (LSTM) networks for time-series domains and standard Convolutional Neural Network (CNN) used for classification. The findings demonstrate that many of the deployed deep learning and machine learning techniques are surpassed by the proposed model. TRANSLATE with x English ArabicHebrewPolishBulgarianHindiPortugueseCatalanHmong DawRomanianChinese SimplifiedHungarianRussianChinese TraditionalIndonesianSlovakCzechItalianSlovenianDanishJapaneseSpanishDutchKlingonSwedishEnglishKoreanThaiEstonianLatvianTurkishFinnishLithuanianUkrainianFrenchMalayUrduGermanMalteseVietnameseGreekNorwegianWelshHaitian CreolePersian // TRANSLATE with COPY THE URL BELOW Back EMBED THE SNIPPET BELOW IN YOUR SITE Enable collaborative features and customize widget: Bing Webmaster Portal Back /
Advanced Map Matching Technologies and Techniques for Pedestrian/Wheelchair Navigation
Due to the constantly increasing technical advantages of mobile devices (such as smartphones), pedestrian/wheelchair navigation recently has achieved a high level of interest as one of smartphones’ potential mobile applications. While vehicle navigation systems have already reached a certain level of maturity, pedestrian/wheelchair navigation services are still in their infancy. By comparing vehicle navigation systems, a set of map matching requirements and challenges unique in pedestrian/wheelchair navigation is identified. To provide navigation assistance to pedestrians and wheelchair users, there is a need for the design and development of new map matching techniques.
The main goal of this research is to investigate and develop advanced map matching technologies and techniques particular for pedestrian/wheelchair navigation services. As the first step in map matching, an adaptive candidate segment selection algorithm is developed to efficiently find candidate segments. Furthermore, to narrow down the search for the correct segment, advanced mathematical models are applied. GPS-based chain-code map matching, Hidden Markov Model (HMM) map matching, and fuzzy-logic map matching algorithms are developed to estimate real-time location of users in pedestrian/wheelchair navigation systems/services. Nevertheless, GPS signal is not always available in areas with high-rise buildings and even when there is a signal, the accuracy may not be high enough for localization of pedestrians and wheelchair users on sidewalks. To overcome these shortcomings of GPS, multi-sensor integrated map matching algorithms are investigated and developed in this research. These algorithms include a movement pattern recognition algorithm, using accelerometer and compass data, and a vision-based positioning algorithm to fill in signal gaps in GPS positioning.
Experiments are conducted to evaluate the developed algorithms using real field test data (GPS coordinates and other sensors data). The experimental results show that the developed algorithms and the integrated sensors, i.e., a monocular visual odometry, a GPS, an accelerometer, and a compass, can provide high-quality and uninterrupted localization services in pedestrian/wheelchair navigation systems/services. The map matching techniques developed in this work can be applied to various pedestrian/wheelchair navigation applications, such as tracking senior citizens and children, or tourist service systems, and can be further utilized in building walking robots and automatic wheelchair navigation systems
Edge Intelligence for Empowering IoT-based Healthcare Systems
The demand for real-time, affordable, and efficient smart healthcare services
is increasing exponentially due to the technological revolution and burst of
population. To meet the increasing demands on this critical infrastructure,
there is a need for intelligent methods to cope with the existing obstacles in
this area. In this regard, edge computing technology can reduce latency and
energy consumption by moving processes closer to the data sources in comparison
to the traditional centralized cloud and IoT-based healthcare systems. In
addition, by bringing automated insights into the smart healthcare systems,
artificial intelligence (AI) provides the possibility of detecting and
predicting high-risk diseases in advance, decreasing medical costs for
patients, and offering efficient treatments. The objective of this article is
to highlight the benefits of the adoption of edge intelligent technology, along
with AI in smart healthcare systems. Moreover, a novel smart healthcare model
is proposed to boost the utilization of AI and edge technology in smart
healthcare systems. Additionally, the paper discusses issues and research
directions arising when integrating these different technologies together.Comment: This paper has been accepted in IEEE Wireless Communication Magazin
A Role-Based Approach for Orchestrating Emergent Configurations in the Internet of Things
The Internet of Things (IoT) is envisioned as a global network of connected
things enabling ubiquitous machine-to-machine (M2M) communication. With
estimations of billions of sensors and devices to be connected in the coming
years, the IoT has been advocated as having a great potential to impact the way
we live, but also how we work. However, the connectivity aspect in itself only
accounts for the underlying M2M infrastructure. In order to properly support
engineering IoT systems and applications, it is key to orchestrate
heterogeneous 'things' in a seamless, adaptive and dynamic manner, such that
the system can exhibit a goal-directed behaviour and take appropriate actions.
Yet, this form of interaction between things needs to take a user-centric
approach and by no means elude the users' requirements. To this end,
contextualisation is an important feature of the system, allowing it to infer
user activities and prompt the user with relevant information and interactions
even in the absence of intentional commands. In this work we propose a
role-based model for emergent configurations of connected systems as a means to
model, manage, and reason about IoT systems including the user's interaction
with them. We put a special focus on integrating the user perspective in order
to guide the emergent configurations such that systems goals are aligned with
the users' intentions. We discuss related scientific and technical challenges
and provide several uses cases outlining the concept of emergent
configurations.Comment: In Proceedings of the Second International Workshop on the Internet
of Agents @AAMAS201
MeciFace: Mechanomyography and Inertial Fusion based Glasses for Edge Real-Time Recognition of Facial and Eating Activities
The increasing prevalence of stress-related eating behaviors and their impact
on overall health highlights the importance of effective monitoring systems. In
this paper, we present MeciFace, an innovative wearable technology designed to
monitor facial expressions and eating activities in real-time on-the-edge
(RTE). MeciFace aims to provide a low-power, privacy-conscious, and highly
accurate tool for promoting healthy eating behaviors and stress management. We
employ lightweight convolutional neural networks as backbone models for facial
expression and eating monitoring scenarios. The MeciFace system ensures
efficient data processing with a tiny memory footprint, ranging from 11KB to
19KB. During RTE evaluation, the system achieves impressive performance,
yielding an F1-score of < 86% for facial expression recognition and 90% for
eating/drinking monitoring, even for the RTE of an unseen user.Comment: Submitted to Nature Scientific Report
Game Theory Solutions in Sensor-Based Human Activity Recognition: A Review
The Human Activity Recognition (HAR) tasks automatically identify human
activities using the sensor data, which has numerous applications in
healthcare, sports, security, and human-computer interaction. Despite
significant advances in HAR, critical challenges still exist. Game theory has
emerged as a promising solution to address these challenges in machine learning
problems including HAR. However, there is a lack of research work on applying
game theory solutions to the HAR problems. This review paper explores the
potential of game theory as a solution for HAR tasks, and bridges the gap
between game theory and HAR research work by suggesting novel game-theoretic
approaches for HAR problems. The contributions of this work include exploring
how game theory can improve the accuracy and robustness of HAR models,
investigating how game-theoretic concepts can optimize recognition algorithms,
and discussing the game-theoretic approaches against the existing HAR methods.
The objective is to provide insights into the potential of game theory as a
solution for sensor-based HAR, and contribute to develop a more accurate and
efficient recognition system in the future research directions
- …