9,939 research outputs found
DeepWalking: Enabling Smartphone-based Walking Speed Estimation Using Deep Learning
Walking speed estimation is an essential component of mobile apps in various
fields such as fitness, transportation, navigation, and health-care. Most
existing solutions are focused on specialized medical applications that utilize
body-worn motion sensors. These approaches do not serve effectively the general
use case of numerous apps where the user holding a smartphone tries to find his
or her walking speed solely based on smartphone sensors. However, existing
smartphone-based approaches fail to provide acceptable precision for walking
speed estimation. This leads to a question: is it possible to achieve
comparable speed estimation accuracy using a smartphone over wearable sensor
based obtrusive solutions?
We find the answer from advanced neural networks. In this paper, we present
DeepWalking, the first deep learning-based walking speed estimation scheme for
smartphone. A deep convolutional neural network (DCNN) is applied to
automatically identify and extract the most effective features from the
accelerometer and gyroscope data of smartphone and to train the network model
for accurate speed estimation. Experiments are performed with 10 participants
using a treadmill. The average root-mean-squared-error (RMSE) of estimated
walking speed is 0.16m/s which is comparable to the results obtained by
state-of-the-art approaches based on a number of body-worn sensors (i.e., RMSE
of 0.11m/s). The results indicate that a smartphone can be a strong tool for
walking speed estimation if the sensor data are effectively calibrated and
supported by advanced deep learning techniques.Comment: 6 pages, 9 figures, published in IEEE Global Communications
Conference (GLOBECOM
Towards a Practical Pedestrian Distraction Detection Framework using Wearables
Pedestrian safety continues to be a significant concern in urban communities
and pedestrian distraction is emerging as one of the main causes of grave and
fatal accidents involving pedestrians. The advent of sophisticated mobile and
wearable devices, equipped with high-precision on-board sensors capable of
measuring fine-grained user movements and context, provides a tremendous
opportunity for designing effective pedestrian safety systems and applications.
Accurate and efficient recognition of pedestrian distractions in real-time
given the memory, computation and communication limitations of these devices,
however, remains the key technical challenge in the design of such systems.
Earlier research efforts in pedestrian distraction detection using data
available from mobile and wearable devices have primarily focused only on
achieving high detection accuracy, resulting in designs that are either
resource intensive and unsuitable for implementation on mainstream mobile
devices, or computationally slow and not useful for real-time pedestrian safety
applications, or require specialized hardware and less likely to be adopted by
most users. In the quest for a pedestrian safety system that achieves a
favorable balance between computational efficiency, detection accuracy, and
energy consumption, this paper makes the following main contributions: (i)
design of a novel complex activity recognition framework which employs motion
data available from users' mobile and wearable devices and a lightweight
frequency matching approach to accurately and efficiently recognize complex
distraction related activities, and (ii) a comprehensive comparative evaluation
of the proposed framework with well-known complex activity recognition
techniques in the literature with the help of data collected from human subject
pedestrians and prototype implementations on commercially-available mobile and
wearable devices
Transportation mode recognition fusing wearable motion, sound and vision sensors
We present the first work that investigates the potential of improving the performance of transportation mode recognition through fusing multimodal data from wearable sensors: motion, sound and vision. We first train three independent deep neural network (DNN) classifiers, which work with the three types of sensors, respectively. We then propose two schemes that fuse the classification results from the three mono-modal classifiers. The first scheme makes an ensemble decision with fixed rules including Sum, Product, Majority Voting, and Borda Count. The second scheme is an adaptive fuser built as another classifier (including Naive Bayes, Decision Tree, Random Forest and Neural Network) that learns enhanced predictions by combining the outputs from the three mono-modal classifiers. We verify the advantage of the proposed method with the state-of-the-art Sussex-Huawei Locomotion and Transportation (SHL) dataset recognizing the eight transportation activities: Still, Walk, Run, Bike, Bus, Car, Train and Subway. We achieve F1 scores of 79.4%, 82.1% and 72.8% with the mono-modal motion, sound and vision classifiers, respectively. The F1 score is remarkably improved to 94.5% and 95.5% by the two data fusion schemes, respectively. The recognition performance can be further improved with a post-processing scheme that exploits the temporal continuity of transportation. When assessing generalization of the model to unseen data, we show that while performance is reduced - as expected - for each individual classifier, the benefits of fusion are retained with performance improved by 15 percentage points. Besides the actual performance increase, this work, most importantly, opens up the possibility for dynamically fusing modalities to achieve distinct power-performance trade-off at run time
Pre-Interaction Identification by Dynamic Grip Classification
We present a novel authentication method to identify users as they pick up a mobile device. We use a combination of back-of-device capacitive sensing and accelerometer measurements to perform classification, and obtain increased performance compared to previous accelerometer-only approaches. Our initial results suggest that users can be reliably identified during the pick-up movement before interaction commences
- …