18,812 research outputs found
Efficient Data Compression with Error Bound Guarantee in Wireless Sensor Networks
We present a data compression and dimensionality reduction scheme for data
fusion and aggregation applications to prevent data congestion and reduce
energy consumption at network connecting points such as cluster heads and
gateways. Our in-network approach can be easily tuned to analyze the data
temporal or spatial correlation using an unsupervised neural network scheme,
namely the autoencoders. In particular, our algorithm extracts intrinsic data
features from previously collected historical samples to transform the raw data
into a low dimensional representation. Moreover, the proposed framework
provides an error bound guarantee mechanism. We evaluate the proposed solution
using real-world data sets and compare it with traditional methods for temporal
and spatial data compression. The experimental validation reveals that our
approach outperforms several existing wireless sensor network's data
compression methods in terms of compression efficiency and signal
reconstruction.Comment: ACM MSWiM 201
Compressive Sampling for Remote Control Systems
In remote control, efficient compression or representation of control signals
is essential to send them through rate-limited channels. For this purpose, we
propose an approach of sparse control signal representation using the
compressive sampling technique. The problem of obtaining sparse representation
is formulated by cardinality-constrained L2 optimization of the control
performance, which is reducible to L1-L2 optimization. The low rate random
sampling employed in the proposed method based on the compressive sampling, in
addition to the fact that the L1-L2 optimization can be effectively solved by a
fast iteration method, enables us to generate the sparse control signal with
reduced computational complexity, which is preferable in remote control systems
where computation delays seriously degrade the performance. We give a
theoretical result for control performance analysis based on the notion of
restricted isometry property (RIP). An example is shown to illustrate the
effectiveness of the proposed approach via numerical experiments
Matching pursuit-based compressive sensing in a wearable biomedical accelerometer fall diagnosis device
There is a significant high fall risk population, where individuals are susceptible to frequent falls and obtaining significant injury, where quick medical response and fall information are critical to providing efficient aid. This article presents an evaluation of compressive sensing techniques in an accelerometer-based intelligent fall detection system modelled on a wearable Shimmer biomedical embedded computing device with Matlab. The presented fall detection system utilises a database of fall and activities of daily living signals evaluated with discrete wavelet transforms and principal component analysis to obtain binary tree classifiers for fall evaluation. 14 test subjects undertook various fall and activities of daily living experiments with a Shimmer device to generate data for principal component analysis-based fall classifiers and evaluate the proposed fall analysis system. The presented system obtains highly accurate fall detection results, demonstrating significant advantages in comparison with the thresholding method presented. Additionally, the presented approach offers advantageous fall diagnostic information. Furthermore, transmitted data accounts for over 80% battery current usage of the Shimmer device, hence it is critical the acceleration data is reduced to increase transmission efficiency and in-turn improve battery usage performance. Various Matching pursuit-based compressive sensing techniques have been utilised to significantly reduce acceleration information required for transmission.Scopu
Improving Performance of Iterative Methods by Lossy Checkponting
Iterative methods are commonly used approaches to solve large, sparse linear
systems, which are fundamental operations for many modern scientific
simulations. When the large-scale iterative methods are running with a large
number of ranks in parallel, they have to checkpoint the dynamic variables
periodically in case of unavoidable fail-stop errors, requiring fast I/O
systems and large storage space. To this end, significantly reducing the
checkpointing overhead is critical to improving the overall performance of
iterative methods. Our contribution is fourfold. (1) We propose a novel lossy
checkpointing scheme that can significantly improve the checkpointing
performance of iterative methods by leveraging lossy compressors. (2) We
formulate a lossy checkpointing performance model and derive theoretically an
upper bound for the extra number of iterations caused by the distortion of data
in lossy checkpoints, in order to guarantee the performance improvement under
the lossy checkpointing scheme. (3) We analyze the impact of lossy
checkpointing (i.e., extra number of iterations caused by lossy checkpointing
files) for multiple types of iterative methods. (4)We evaluate the lossy
checkpointing scheme with optimal checkpointing intervals on a high-performance
computing environment with 2,048 cores, using a well-known scientific
computation package PETSc and a state-of-the-art checkpoint/restart toolkit.
Experiments show that our optimized lossy checkpointing scheme can
significantly reduce the fault tolerance overhead for iterative methods by
23%~70% compared with traditional checkpointing and 20%~58% compared with
lossless-compressed checkpointing, in the presence of system failures.Comment: 14 pages, 10 figures, HPDC'1
- …