3,657 research outputs found
Compression via Compressive Sensing : A Low-Power Framework for the Telemonitoring of Multi-Channel Physiological Signals
Telehealth and wearable equipment can deliver personal healthcare and
necessary treatment remotely. One major challenge is transmitting large amount
of biosignals through wireless networks. The limited battery life calls for
low-power data compressors. Compressive Sensing (CS) has proved to be a
low-power compressor. In this study, we apply CS on the compression of
multichannel biosignals. We firstly develop an efficient CS algorithm from the
Block Sparse Bayesian Learning (BSBL) framework. It is based on a combination
of the block sparse model and multiple measurement vector model. Experiments on
real-life Fetal ECGs showed that the proposed algorithm has high fidelity and
efficiency. Implemented in hardware, the proposed algorithm was compared to a
Discrete Wavelet Transform (DWT) based algorithm, verifying the proposed one
has low power consumption and occupies less computational resources.Comment: 2013 International Workshop on Biomedical and Health Informatic
Hardware-Efficient Structure of the Accelerating Module for Implementation of Convolutional Neural Network Basic Operation
This paper presents a structural design of the hardware-efficient module for
implementation of convolution neural network (CNN) basic operation with reduced
implementation complexity. For this purpose we utilize some modification of the
Winograd minimal filtering method as well as computation vectorization
principles. This module calculate inner products of two consecutive segments of
the original data sequence, formed by a sliding window of length 3, with the
elements of a filter impulse response. The fully parallel structure of the
module for calculating these two inner products, based on the implementation of
a naive method of calculation, requires 6 binary multipliers and 4 binary
adders. The use of the Winograd minimal filtering method allows to construct a
module structure that requires only 4 binary multipliers and 8 binary adders.
Since a high-performance convolutional neural network can contain tens or even
hundreds of such modules, such a reduction can have a significant effect.Comment: 3 pages, 5 figure
Optimising Sparse Matrix Vector multiplication for large scale FEM problems on FPGA
Sparse Matrix Vector multiplication (SpMV) is an important kernel in many scientific applications. In this work we propose an architecture and an automated customisation method to detect and optimise the architecture for block diagonal sparse matrices. We evaluate the proposed approach in the context of the spectral/hp Finite Element Method, using the local matrix assembly approach. This problem leads to a large sparse system of linear equations with block diagonal matrix which is typically solved using an iterative method such as the Preconditioned Conjugate Gradient. The efficiency of the proposed architecture combined with the effectiveness of the proposed customisation method reduces BRAM resource utilisation by as much as 10 times, while achieving identical throughput with existing state of the art designs and requiring minimal development effort from the end user. In the context of the Finite Element Method, our approach enables the solution of larger problems than previously possible, enabling the applicability of FPGAs to more interesting HPC problems
- …