93 research outputs found

    Compressively Sensed Image Recognition

    Full text link
    Compressive Sensing (CS) theory asserts that sparse signal reconstruction is possible from a small number of linear measurements. Although CS enables low-cost linear sampling, it requires non-linear and costly reconstruction. Recent literature works show that compressive image classification is possible in CS domain without reconstruction of the signal. In this work, we introduce a DCT base method that extracts binary discriminative features directly from CS measurements. These CS measurements can be obtained by using (i) a random or a pseudo-random measurement matrix, or (ii) a measurement matrix whose elements are learned from the training data to optimize the given classification task. We further introduce feature fusion by concatenating Bag of Words (BoW) representation of our binary features with one of the two state-of-the-art CNN-based feature vectors. We show that our fused feature outperforms the state-of-the-art in both cases.Comment: 6 pages, submitted/accepted, EUVIP 201

    Improving Maternal and Fetal Cardiac Monitoring Using Artificial Intelligence

    Get PDF
    Early diagnosis of possible risks in the physiological status of fetus and mother during pregnancy and delivery is critical and can reduce mortality and morbidity. For example, early detection of life-threatening congenital heart disease may increase survival rate and reduce morbidity while allowing parents to make informed decisions. To study cardiac function, a variety of signals are required to be collected. In practice, several heart monitoring methods, such as electrocardiogram (ECG) and photoplethysmography (PPG), are commonly performed. Although there are several methods for monitoring fetal and maternal health, research is currently underway to enhance the mobility, accuracy, automation, and noise resistance of these methods to be used extensively, even at home. Artificial Intelligence (AI) can help to design a precise and convenient monitoring system. To achieve the goals, the following objectives are defined in this research: The first step for a signal acquisition system is to obtain high-quality signals. As the first objective, a signal processing scheme is explored to improve the signal-to-noise ratio (SNR) of signals and extract the desired signal from a noisy one with negative SNR (i.e., power of noise is greater than signal). It is worth mentioning that ECG and PPG signals are sensitive to noise from a variety of sources, increasing the risk of misunderstanding and interfering with the diagnostic process. The noises typically arise from power line interference, white noise, electrode contact noise, muscle contraction, baseline wandering, instrument noise, motion artifacts, electrosurgical noise. Even a slight variation in the obtained ECG waveform can impair the understanding of the patient's heart condition and affect the treatment procedure. Recent solutions, such as adaptive and blind source separation (BSS) algorithms, still have drawbacks, such as the need for noise or desired signal model, tuning and calibration, and inefficiency when dealing with excessively noisy signals. Therefore, the final goal of this step is to develop a robust algorithm that can estimate noise, even when SNR is negative, using the BSS method and remove it based on an adaptive filter. The second objective is defined for monitoring maternal and fetal ECG. Previous methods that were non-invasive used maternal abdominal ECG (MECG) for extracting fetal ECG (FECG). These methods need to be calibrated to generalize well. In other words, for each new subject, a calibration with a trustable device is required, which makes it difficult and time-consuming. The calibration is also susceptible to errors. We explore deep learning (DL) models for domain mapping, such as Cycle-Consistent Adversarial Networks, to map MECG to fetal ECG (FECG) and vice versa. The advantages of the proposed DL method over state-of-the-art approaches, such as adaptive filters or blind source separation, are that the proposed method is generalized well on unseen subjects. Moreover, it does not need calibration and is not sensitive to the heart rate variability of mother and fetal; it can also handle low signal-to-noise ratio (SNR) conditions. Thirdly, AI-based system that can measure continuous systolic blood pressure (SBP) and diastolic blood pressure (DBP) with minimum electrode requirements is explored. The most common method of measuring blood pressure is using cuff-based equipment, which cannot monitor blood pressure continuously, requires calibration, and is difficult to use. Other solutions use a synchronized ECG and PPG combination, which is still inconvenient and challenging to synchronize. The proposed method overcomes those issues and only uses PPG signal, comparing to other solutions. Using only PPG for blood pressure is more convenient since it is only one electrode on the finger where its acquisition is more resilient against error due to movement. The fourth objective is to detect anomalies on FECG data. The requirement of thousands of manually annotated samples is a concern for state-of-the-art detection systems, especially for fetal ECG (FECG), where there are few publicly available FECG datasets annotated for each FECG beat. Therefore, we will utilize active learning and transfer-learning concept to train a FECG anomaly detection system with the least training samples and high accuracy. In this part, a model is trained for detecting ECG anomalies in adults. Later this model is trained to detect anomalies on FECG. We only select more influential samples from the training set for training, which leads to training with the least effort. Because of physician shortages and rural geography, pregnant women's ability to get prenatal care might be improved through remote monitoring, especially when access to prenatal care is limited. Increased compliance with prenatal treatment and linked care amongst various providers are two possible benefits of remote monitoring. If recorded signals are transmitted correctly, maternal and fetal remote monitoring can be effective. Therefore, the last objective is to design a compression algorithm that can compress signals (like ECG) with a higher ratio than state-of-the-art and perform decompression fast without distortion. The proposed compression is fast thanks to the time domain B-Spline approach, and compressed data can be used for visualization and monitoring without decompression owing to the B-spline properties. Moreover, the stochastic optimization is designed to retain the signal quality and does not distort signal for diagnosis purposes while having a high compression ratio. In summary, components for creating an end-to-end system for day-to-day maternal and fetal cardiac monitoring can be envisioned as a mix of all tasks listed above. PPG and ECG recorded from the mother can be denoised using deconvolution strategy. Then, compression can be employed for transmitting signal. The trained CycleGAN model can be used for extracting FECG from MECG. Then, trained model using active transfer learning can detect anomaly on both MECG and FECG. Simultaneously, maternal BP is retrieved from the PPG signal. This information can be used for monitoring the cardiac status of mother and fetus, and also can be used for filling reports such as partogram

    Sensor Signal and Information Processing II

    Get PDF
    In the current age of information explosion, newly invented technological sensors and software are now tightly integrated with our everyday lives. Many sensor processing algorithms have incorporated some forms of computational intelligence as part of their core framework in problem solving. These algorithms have the capacity to generalize and discover knowledge for themselves and learn new information whenever unseen data are captured. The primary aim of sensor processing is to develop techniques to interpret, understand, and act on information contained in the data. The interest of this book is in developing intelligent signal processing in order to pave the way for smart sensors. This involves mathematical advancement of nonlinear signal processing theory and its applications that extend far beyond traditional techniques. It bridges the boundary between theory and application, developing novel theoretically inspired methodologies targeting both longstanding and emergent signal processing applications. The topic ranges from phishing detection to integration of terrestrial laser scanning, and from fault diagnosis to bio-inspiring filtering. The book will appeal to established practitioners, along with researchers and students in the emerging field of smart sensors processing

    Efficient ECG Compression and QRS Detection for E-Health Applications

    Get PDF
    Current medical screening and diagnostic procedures have shifted toward recording longer electrocardiogram (ECG) signals, which have traditionally been processed on personal computers (PCs) with high-speed multi-core processors and efficient memory processing. Battery-driven devices are now more commonly used for the same purpose and thus exploring highly efficient, low-power alternatives for local ECG signal collection and processing is essential for efficient and convenient clinical use. Several ECG compression methods have been reported in the current literature with limited discussion on the performance of the compressed and the reconstructed ECG signals in terms of the QRS complex detection accuracy. This paper proposes and evaluates different compression methods based not only on the compression ratio (CR) and percentage root-mean-square difference (PRD), but also based on the accuracy of QRS detection. In this paper, we have developed a lossy method (Methods III) and compared them to the most current lossless and lossy ECG compression methods (Method I and Method II, respectively). The proposed lossy compression method (Method III) achieves CR of 4.5×, PRD of 0.53, as well as an overall sensitivity of 99.78% and positive predictivity of 99.92% are achieved (when coupled with an existing QRS detection algorithm) on the MIT-BIH Arrhythmia database and an overall sensitivity of 99.90% and positive predictivity of 99.84% on the QT database.This work was made possible by NPRP grant #7-684-1-127 from the Qatar National Research Fund (a member of Qatar Foundation). The statements made herein are solely the responsibility of the authors.Scopu

    Algorithms design for improving homecare using Electrocardiogram (ECG) signals and Internet of Things (IoT)

    Get PDF
    Due to the fast growing of population, a lot of hospitals get crowded from the huge amount of patients visits. Moreover, during COVID-19 a lot of patients prefer staying at home to minimize the spread of the virus. The need for providing care to patients at home is essential. Internet of Things (IoT) is widely known and used by different fields. IoT based homecare will help in reducing the burden upon hospitals. IoT with homecare bring up several benefits such as minimizing human exertions, economical savings and improved efficiency and effectiveness. One of the important requirement on homecare system is the accuracy because those systems are dealing with human health which is sensitive and need high amount of accuracy. Moreover, those systems deal with huge amount of data due to the continues sensing that need to be processed well to provide fast response regarding the diagnosis with minimum cost requirements. Heart is one of the most important organ in the human body that requires high level of caring. Monitoring heart status can diagnose disease from the early stage and find the best medication plan by health experts. Continues monitoring and diagnosis of heart could exhaust caregivers efforts. Having an IoT heart monitoring model at home is the solution to this problem. Electrocardiogram (ECG) signals are used to track heart condition using waves and peaks. Accurate and efficient IoT ECG monitoring at home can detect heart diseases and save human lives. As a consequence, an IoT ECG homecare monitoring model is designed in this thesis for detecting Cardiac Arrhythmia and diagnosing heart diseases. Two databases of ECG signals are used; one online which is old and limited, and another huge, unique and special from real patients in hospital. The raw ECG signal for each patient is passed through the implemented Low Pass filter and Savitzky Golay filter signal processing techniques to remove the noise and any external interference. The clear signal in this model is passed through feature extraction stage to extract number of features based on some metrics and medical information along with feature extraction algorithm to find peaks and waves. Those features are saved in the local database to apply classification on them. For the diagnosis purpose a classification stage is made using three classification ways; threshold values, machine learning and deep learning to increase the accuracy. Threshold values classification technique worked based on medical values and boarder lines. In case any feature goes above or beyond these ranges, a warning message appeared with expected heart disease. The second type of classification is by using machine learning to minimize the human efforts. A Support Vector Machine (SVM) algorithm is proposed by running the algorithm on the features extracted from both databases. The classification accuracy for online and hospital databases was 91.67% and 94% respectively. Due to the non-linearity of the decision boundary, a third way of classification using deep learning is presented. A full Multilayer Perceptron (MLP) Neural Network is implemented to improve the accuracy and reduce the errors. The number of errors reduced to 0.019 and 0.006 using online and hospital databases. While using hospital database which is huge, there is a need for a technique to reduce the amount of data. Furthermore, a novel adaptive amplitude threshold compression algorithm is proposed. This algorithm is able to make diagnosis of heart disease from the reduced size using compressed ECG signals with high level of accuracy and low cost. The extracted features from compressed and original are similar with only slight differences of 1%, 2% and 3% with no effects on machine learning and deep learning classification accuracy without the need for any reconstructions. The throughput is improved by 43% with reduced storage space of 57% when using data compression. Moreover, to achieve fast response, the amount of data should be reduced further to provide fast data transmission. A compressive sensing based cardiac homecare system is presented. It gives the channel between sender and receiver the ability to carry small amount of data. Experiment results reveal that the proposed models are more accurate in the classification of Cardiac Arrhythmia and in the diagnosis of heart diseases. The proposed models ensure fast diagnosis and minimum cost requirements. Based on the experiments on classification accuracy, number of errors and false alarms, the dictionary of the compressive sensing selected to be 900. As a result, this thesis provided three different scenarios that achieved IoT homecare Cardiac monitoring to assist in further research for designing homecare Cardiac monitoring systems. The experiment results reveal that those scenarios produced better results with high level of accuracy in addition to minimizing data and cost requirements

    Sparse signal representation, sampling, and recovery in compressive sensing frameworks

    Get PDF
    Compressive sensing allows the reconstruction of original signals from a much smaller number of samples as compared to the Nyquist sampling rate. The effectiveness of compressive sensing motivated the researchers for its deployment in a variety of application areas. The use of an efficient sampling matrix for high-performance recovery algorithms improves the performance of the compressive sensing framework significantly. This paper presents the underlying concepts of compressive sensing as well as previous work done in targeted domains in accordance with the various application areas. To develop prospects within the available functional blocks of compressive sensing frameworks, a diverse range of application areas are investigated. The three fundamental elements of a compressive sensing framework (signal sparsity, subsampling, and reconstruction) are thoroughly reviewed in this work by becoming acquainted with the key research gaps previously identified by the research community. Similarly, the basic mathematical formulation is used to outline some primary performance evaluation metrics for 1D and 2D compressive sensing.Web of Science10850188500

    SPARSE RECOVERY BY NONCONVEX LIPSHITZIAN MAPPINGS

    Get PDF
    In recent years, the sparsity concept has attracted considerable attention in areas of applied mathematics and computer science, especially in signal and image processing fields. The general framework of sparse representation is now a mature concept with solid basis in relevant mathematical fields, such as probability, geometry of Banach spaces, harmonic analysis, theory of computability, and information-based complexity. Together with theoretical and practical advancements, also several numeric methods and algorithmic techniques have been developed in order to capture the complexity and the wide scope that the theory suggests. Sparse recovery relays over the fact that many signals can be represented in a sparse way, using only few nonzero coefficients in a suitable basis or overcomplete dictionary. Unfortunately, this problem, also called `0-norm minimization, is not only NP-hard, but also hard to approximate within an exponential factor of the optimal solution. Nevertheless, many heuristics for the problem has been obtained and proposed for many applications. This thesis provides new regularization methods for the sparse representation problem with application to face recognition and ECG signal compression. The proposed methods are based on fixed-point iteration scheme which combines nonconvex Lipschitzian-type mappings with canonical orthogonal projectors. The first are aimed at uniformly enhancing the sparseness level by shrinking effects, the latter to project back into the feasible space of solutions. In the second part of this thesis we study two applications in which sparseness has been successfully applied in recent areas of the signal and image processing: the face recognition problem and the ECG signal compression problem
    • …
    corecore