3,082 research outputs found
DeepWalking: Enabling Smartphone-based Walking Speed Estimation Using Deep Learning
Walking speed estimation is an essential component of mobile apps in various
fields such as fitness, transportation, navigation, and health-care. Most
existing solutions are focused on specialized medical applications that utilize
body-worn motion sensors. These approaches do not serve effectively the general
use case of numerous apps where the user holding a smartphone tries to find his
or her walking speed solely based on smartphone sensors. However, existing
smartphone-based approaches fail to provide acceptable precision for walking
speed estimation. This leads to a question: is it possible to achieve
comparable speed estimation accuracy using a smartphone over wearable sensor
based obtrusive solutions?
We find the answer from advanced neural networks. In this paper, we present
DeepWalking, the first deep learning-based walking speed estimation scheme for
smartphone. A deep convolutional neural network (DCNN) is applied to
automatically identify and extract the most effective features from the
accelerometer and gyroscope data of smartphone and to train the network model
for accurate speed estimation. Experiments are performed with 10 participants
using a treadmill. The average root-mean-squared-error (RMSE) of estimated
walking speed is 0.16m/s which is comparable to the results obtained by
state-of-the-art approaches based on a number of body-worn sensors (i.e., RMSE
of 0.11m/s). The results indicate that a smartphone can be a strong tool for
walking speed estimation if the sensor data are effectively calibrated and
supported by advanced deep learning techniques.Comment: 6 pages, 9 figures, published in IEEE Global Communications
Conference (GLOBECOM
Vanishing Point Detection with Direct and Transposed Fast Hough Transform inside the neural network
In this paper, we suggest a new neural network architecture for vanishing
point detection in images. The key element is the use of the direct and
transposed Fast Hough Transforms separated by convolutional layer blocks with
standard activation functions. It allows us to get the answer in the
coordinates of the input image at the output of the network and thus to
calculate the coordinates of the vanishing point by simply selecting the
maximum. Besides, it was proved that calculation of the transposed Fast Hough
Transform can be performed using the direct one. The use of integral operators
enables the neural network to rely on global rectilinear features in the image,
and so it is ideal for detecting vanishing points. To demonstrate the
effectiveness of the proposed architecture, we use a set of images from a DVR
and show its superiority over existing methods. Note, in addition, that the
proposed neural network architecture essentially repeats the process of direct
and back projection used, for example, in computed tomography.Comment: 9 pages, 9 figures, submitted to "Computer Optics"; extra experiment
added, new theorem proof added, references added; typos correcte
Anytime Stereo Image Depth Estimation on Mobile Devices
Many applications of stereo depth estimation in robotics require the
generation of accurate disparity maps in real time under significant
computational constraints. Current state-of-the-art algorithms force a choice
between either generating accurate mappings at a slow pace, or quickly
generating inaccurate ones, and additionally these methods typically require
far too many parameters to be usable on power- or memory-constrained devices.
Motivated by these shortcomings, we propose a novel approach for disparity
prediction in the anytime setting. In contrast to prior work, our end-to-end
learned approach can trade off computation and accuracy at inference time.
Depth estimation is performed in stages, during which the model can be queried
at any time to output its current best estimate. Our final model can process
1242375 resolution images within a range of 10-35 FPS on an NVIDIA
Jetson TX2 module with only marginal increases in error -- using two orders of
magnitude fewer parameters than the most competitive baseline. The source code
is available at https://github.com/mileyan/AnyNet .Comment: Accepted by ICRA201
DPC-Net: Deep Pose Correction for Visual Localization
We present a novel method to fuse the power of deep networks with the
computational efficiency of geometric and probabilistic localization
algorithms. In contrast to other methods that completely replace a classical
visual estimator with a deep network, we propose an approach that uses a
convolutional neural network to learn difficult-to-model corrections to the
estimator from ground-truth training data. To this end, we derive a novel loss
function for learning SE(3) corrections based on a matrix Lie groups approach,
with a natural formulation for balancing translation and rotation errors. We
use this loss to train a Deep Pose Correction network (DPC-Net) that predicts
corrections for a particular estimator, sensor and environment. Using the KITTI
odometry dataset, we demonstrate significant improvements to the accuracy of a
computationally-efficient sparse stereo visual odometry pipeline, that render
it as accurate as a modern computationally-intensive dense estimator. Further,
we show how DPC-Net can be used to mitigate the effect of poorly calibrated
lens distortion parameters.Comment: In IEEE Robotics and Automation Letters (RA-L) and presented at the
IEEE International Conference on Robotics and Automation (ICRA'18), Brisbane,
Australia, May 21-25, 201
- …