1,597 research outputs found

    Compressively characterizing high-dimensional entangled states with complementary, random filtering

    Get PDF
    The resources needed to conventionally characterize a quantum system are overwhelmingly large for high- dimensional systems. This obstacle may be overcome by abandoning traditional cornerstones of quantum measurement, such as general quantum states, strong projective measurement, and assumption-free characterization. Following this reasoning, we demonstrate an efficient technique for characterizing high-dimensional, spatial entanglement with one set of measurements. We recover sharp distributions with local, random filtering of the same ensemble in momentum followed by position---something the uncertainty principle forbids for projective measurements. Exploiting the expectation that entangled signals are highly correlated, we use fewer than 5,000 measurements to characterize a 65, 536-dimensional state. Finally, we use entropic inequalities to witness entanglement without a density matrix. Our method represents the sea change unfolding in quantum measurement where methods influenced by the information theory and signal-processing communities replace unscalable, brute-force techniques---a progression previously followed by classical sensing.Comment: 13 pages, 7 figure

    Robust and Efficient Inference of Scene and Object Motion in Multi-Camera Systems

    Get PDF
    Multi-camera systems have the ability to overcome some of the fundamental limitations of single camera based systems. Having multiple view points of a scene goes a long way in limiting the influence of field of view, occlusion, blur and poor resolution of an individual camera. This dissertation addresses robust and efficient inference of object motion and scene in multi-camera and multi-sensor systems. The first part of the dissertation discusses the role of constraints introduced by projective imaging towards robust inference of multi-camera/sensor based object motion. We discuss the role of the homography and epipolar constraints for fusing object motion perceived by individual cameras. For planar scenes, the homography constraints provide a natural mechanism for data association. For scenes that are not planar, the epipolar constraint provides a weaker multi-view relationship. We use the epipolar constraint for tracking in multi-camera and multi-sensor networks. In particular, we show that the epipolar constraint reduces the dimensionality of the state space of the problem by introducing a ``shared'' state space for the joint tracking problem. This allows for robust tracking even when one of the sensors fail due to poor SNR or occlusion. The second part of the dissertation deals with challenges in the computational aspects of tracking algorithms that are common to such systems. Much of the inference in the multi-camera and multi-sensor networks deal with complex non-linear models corrupted with non-Gaussian noise. Particle filters provide approximate Bayesian inference in such settings. We analyze the computational drawbacks of traditional particle filtering algorithms, and present a method for implementing the particle filter using the Independent Metropolis Hastings sampler, that is highly amenable to pipelined implementations and parallelization. We analyze the implementations of the proposed algorithm, and in particular concentrate on implementations that have minimum processing times. The last part of the dissertation deals with the efficient sensing paradigm of compressing sensing (CS) applied to signals in imaging, such as natural images and reflectance fields. We propose a hybrid signal model on the assumption that most real-world signals exhibit subspace compressibility as well as sparse representations. We show that several real-world visual signals such as images, reflectance fields, videos etc., are better approximated by this hybrid of two models. We derive optimal hybrid linear projections of the signal and show that theoretical guarantees and algorithms designed for CS can be easily extended to hybrid subspace-compressive sensing. Such methods reduce the amount of information sensed by a camera, and help in reducing the so called data deluge problem in large multi-camera systems

    Ensemble approach on enhanced compressed noise EEG data signal in wireless body area sensor network

    Get PDF
    The Wireless Body Area Sensor Network (WBASN) is used for communication among sensor nodes operating on or inside the human body in order to monitor vital body parameters and movements. One of the important applications of WBASN is patients’ healthcare monitoring of chronic diseases such as epileptic seizure. Normally, epileptic seizure data of the electroencephalograph (EEG) is captured and compressed in order to reduce its transmission time. However, at the same time, this contaminates the overall data and lowers classification accuracy. The current work also did not take into consideration that large size of collected EEG data. Consequently, EEG data is a bandwidth intensive. Hence, the main goal of this work is to design a unified compression and classification framework for delivery of EEG data in order to address its large size issue. EEG data is compressed in order to reduce its transmission time. However, at the same time, noise at the receiver side contaminates the overall data and lowers classification accuracy. Another goal is to reconstruct the compressed data and then recognize it. Therefore, a Noise Signal Combination (NSC) technique is proposed for the compression of the transmitted EEG data and enhancement of its classification accuracy at the receiving side in the presence of noise and incomplete data. The proposed framework combines compressive sensing and discrete cosine transform (DCT) in order to reduce the size of transmission data. Moreover, Gaussian noise model of the transmission channel is practically implemented to the framework. At the receiving side, the proposed NSC is designed based on weighted voting using four classification techniques. The accuracy of these techniques namely Artificial Neural Network, Naïve Bayes, k-Nearest Neighbour, and Support Victor Machine classifiers is fed to the proposed NSC. The experimental results showed that the proposed technique exceeds the conventional techniques by achieving the highest accuracy for noiseless and noisy data. Furthermore, the framework performs a significant role in reducing the size of data and classifying both noisy and noiseless data. The key contributions are the unified framework and proposed NSC, which improved accuracy of the noiseless and noisy EGG large data. The results have demonstrated the effectiveness of the proposed framework and provided several credible benefits including simplicity, and accuracy enhancement. Finally, the research improves clinical information about patients who not only suffer from epilepsy, but also neurological disorders, mental or physiological problems

    Algorithms design for improving homecare using Electrocardiogram (ECG) signals and Internet of Things (IoT)

    Get PDF
    Due to the fast growing of population, a lot of hospitals get crowded from the huge amount of patients visits. Moreover, during COVID-19 a lot of patients prefer staying at home to minimize the spread of the virus. The need for providing care to patients at home is essential. Internet of Things (IoT) is widely known and used by different fields. IoT based homecare will help in reducing the burden upon hospitals. IoT with homecare bring up several benefits such as minimizing human exertions, economical savings and improved efficiency and effectiveness. One of the important requirement on homecare system is the accuracy because those systems are dealing with human health which is sensitive and need high amount of accuracy. Moreover, those systems deal with huge amount of data due to the continues sensing that need to be processed well to provide fast response regarding the diagnosis with minimum cost requirements. Heart is one of the most important organ in the human body that requires high level of caring. Monitoring heart status can diagnose disease from the early stage and find the best medication plan by health experts. Continues monitoring and diagnosis of heart could exhaust caregivers efforts. Having an IoT heart monitoring model at home is the solution to this problem. Electrocardiogram (ECG) signals are used to track heart condition using waves and peaks. Accurate and efficient IoT ECG monitoring at home can detect heart diseases and save human lives. As a consequence, an IoT ECG homecare monitoring model is designed in this thesis for detecting Cardiac Arrhythmia and diagnosing heart diseases. Two databases of ECG signals are used; one online which is old and limited, and another huge, unique and special from real patients in hospital. The raw ECG signal for each patient is passed through the implemented Low Pass filter and Savitzky Golay filter signal processing techniques to remove the noise and any external interference. The clear signal in this model is passed through feature extraction stage to extract number of features based on some metrics and medical information along with feature extraction algorithm to find peaks and waves. Those features are saved in the local database to apply classification on them. For the diagnosis purpose a classification stage is made using three classification ways; threshold values, machine learning and deep learning to increase the accuracy. Threshold values classification technique worked based on medical values and boarder lines. In case any feature goes above or beyond these ranges, a warning message appeared with expected heart disease. The second type of classification is by using machine learning to minimize the human efforts. A Support Vector Machine (SVM) algorithm is proposed by running the algorithm on the features extracted from both databases. The classification accuracy for online and hospital databases was 91.67% and 94% respectively. Due to the non-linearity of the decision boundary, a third way of classification using deep learning is presented. A full Multilayer Perceptron (MLP) Neural Network is implemented to improve the accuracy and reduce the errors. The number of errors reduced to 0.019 and 0.006 using online and hospital databases. While using hospital database which is huge, there is a need for a technique to reduce the amount of data. Furthermore, a novel adaptive amplitude threshold compression algorithm is proposed. This algorithm is able to make diagnosis of heart disease from the reduced size using compressed ECG signals with high level of accuracy and low cost. The extracted features from compressed and original are similar with only slight differences of 1%, 2% and 3% with no effects on machine learning and deep learning classification accuracy without the need for any reconstructions. The throughput is improved by 43% with reduced storage space of 57% when using data compression. Moreover, to achieve fast response, the amount of data should be reduced further to provide fast data transmission. A compressive sensing based cardiac homecare system is presented. It gives the channel between sender and receiver the ability to carry small amount of data. Experiment results reveal that the proposed models are more accurate in the classification of Cardiac Arrhythmia and in the diagnosis of heart diseases. The proposed models ensure fast diagnosis and minimum cost requirements. Based on the experiments on classification accuracy, number of errors and false alarms, the dictionary of the compressive sensing selected to be 900. As a result, this thesis provided three different scenarios that achieved IoT homecare Cardiac monitoring to assist in further research for designing homecare Cardiac monitoring systems. The experiment results reveal that those scenarios produced better results with high level of accuracy in addition to minimizing data and cost requirements

    Statistical Filtering for Multimodal Mobility Modeling in Cyber Physical Systems

    Get PDF
    A Cyber-Physical System integrates computations and dynamics of physical processes. It is an engineering discipline focused on technology with a strong foundation in mathematical abstractions. It shares many of these abstractions with engineering and computer science, but still requires adaptation to suit the dynamics of the physical world. In such a dynamic system, mobility management is one of the key issues against developing a new service. For example, in the study of a new mobile network, it is necessary to simulate and evaluate a protocol before deployment in the system. Mobility models characterize mobile agent movement patterns. On the other hand, they describe the conditions of the mobile services. The focus of this thesis is on mobility modeling in cyber-physical systems. A macroscopic model that captures the mobility of individuals (people and vehicles) can facilitate an unlimited number of applications. One fundamental and obvious example is traffic profiling. Mobility in most systems is a dynamic process and small non-linearities can lead to substantial errors in the model. Extensive research activities on statistical inference and filtering methods for data modeling in cyber-physical systems exist. In this thesis, several methods are employed for multimodal data fusion, localization and traffic modeling. A novel energy-aware sparse signal processing method is presented to process massive sensory data. At baseline, this research examines the application of statistical filters for mobility modeling and assessing the difficulties faced in fusing massive multi-modal sensory data. A statistical framework is developed to apply proposed methods on available measurements in cyber-physical systems. The proposed methods have employed various statistical filtering schemes (i.e., compressive sensing, particle filtering and kernel-based optimization) and applied them to multimodal data sets, acquired from intelligent transportation systems, wireless local area networks, cellular networks and air quality monitoring systems. Experimental results show the capability of these proposed methods in processing multimodal sensory data. It provides a macroscopic mobility model of mobile agents in an energy efficient way using inconsistent measurements

    Proceedings of the second "international Traveling Workshop on Interactions between Sparse models and Technology" (iTWIST'14)

    Get PDF
    The implicit objective of the biennial "international - Traveling Workshop on Interactions between Sparse models and Technology" (iTWIST) is to foster collaboration between international scientific teams by disseminating ideas through both specific oral/poster presentations and free discussions. For its second edition, the iTWIST workshop took place in the medieval and picturesque town of Namur in Belgium, from Wednesday August 27th till Friday August 29th, 2014. The workshop was conveniently located in "The Arsenal" building within walking distance of both hotels and town center. iTWIST'14 has gathered about 70 international participants and has featured 9 invited talks, 10 oral presentations, and 14 posters on the following themes, all related to the theory, application and generalization of the "sparsity paradigm": Sparsity-driven data sensing and processing; Union of low dimensional subspaces; Beyond linear and convex inverse problem; Matrix/manifold/graph sensing/processing; Blind inverse problems and dictionary learning; Sparsity and computational neuroscience; Information theory, geometry and randomness; Complexity/accuracy tradeoffs in numerical methods; Sparsity? What's next?; Sparse machine learning and inference.Comment: 69 pages, 24 extended abstracts, iTWIST'14 website: http://sites.google.com/site/itwist1
    • …
    corecore