12 research outputs found

    Multi-scale Sparse Coding With Anomaly Detection And Classification

    Get PDF
    We here place a recent joint anomaly detection and classification approach based on sparse error coding methodology into multi-scale wavelet basis framework. The model is extended to incorporate an overcomplete wavelet basis into the dictionary matrix whereupon anomalies at specified multiple levels of scale are afforded equal importance. This enables, for example, subtle transient anomalies at finer scales to be detected which would otherwise be drowned out by coarser details and missed by the standard sparse coding techniques. Anomaly detection in power networks provides a motivating application and tests on a real-world data set corroborates the efficacy of the proposed model

    Cardiac Motion Estimation with Dictionary Learning and Robust Sparse Coding in Ultrasound Imaging

    Get PDF
    Cardiac motion estimation from ultrasound images is an ill-posed problem that needs regularization to stabilize the solution. In this work, regularization is achieved by exploiting the sparseness of cardiac motion fields when decomposed in an appropriate dictionary, as well as their smoothness through a classical total variation term. The main contribution of this work is to robustify the sparse coding step in order to handle anomalies, i.e., motion patterns that significantly deviate from the expected model. The proposed approach uses an ADMM-based optimization algorithm in order to simultaneously recover the sparse representations and the outlier components. It is evaluated using two realistic simulated datasets with available ground-truth, containing native outliers and corrupted by synthetic attenuation and clutter artefacts

    Machine Learning Techniques for Personalized Detection of Epileptic Events in Clinical Video Recordings

    Full text link
    Continuous patient monitoring is essential to achieve an effective and optimal patient treatment in the intensive care unit. In the specific case of epilepsy it is the only way to achieve a correct diagnosis and a subsequent optimal medication plan if possible. In addition to automatic vital sign monitoring, epilepsy patients need manual monitoring by trained personnel, a task that is very difficult to be performed continuously for each patient. Moreover, epileptic manifestations are highly personalized even within the same type of epilepsy. In this work we assess two machine learning methods, dictionary learning and an autoencoder based on long short-term memory (LSTM) cells, on the task of personalized epileptic event detection in videos, with a set of features that were specifically developed with an emphasis on high motion sensitivity. According to the strengths of each method we have selected different types of epilepsy, one with convulsive behaviour and one with very subtle motion. The results on five clinical patients show a highly promising ability of both methods to detect the epileptic events as anomalies deviating from the stable/normal patient status

    Fast sparse coding for range data denoising with sparse ridges constraint

    Get PDF
    Light detection and ranging (LiDAR) sensors have been widely deployed on intelligent systems such as unmanned ground vehicles (UGVs) and unmanned aerial vehicles (UAVs) to perform localization, obstacle detection, and navigation tasks. Thus, research into range data processing with competitive performance in terms of both accuracy and efficiency has attracted increasing attention. Sparse coding has revolutionized signal processing and led to state-of-the-art performance in a variety of applications. However, dictionary learning, which plays the central role in sparse coding techniques, is computationally demanding, resulting in its limited applicability in real-time systems. In this study, we propose sparse coding algorithms with a fixed pre-learned ridge dictionary to realize range data denoising via leveraging the regularity of laser range measurements in man-made environments. Experiments on both synthesized data and real data demonstrate that our method obtains accuracy comparable to that of sophisticated sparse coding methods, but with much higher computational efficiency

    Representation Learning via Cauchy Convolutional Sparse Coding

    Get PDF
    In representation learning, Convolutional Sparse Coding (CSC) enables unsupervised learning of features by jointly optimising both an 2\ell_2-norm fidelity term and a sparsity enforcing penalty. This work investigates using a regularisation term derived from an assumed Cauchy prior for the coefficients of the feature maps of a CSC generative model. The sparsity penalty term resulting from this prior is solved via its proximal operator, which is then applied iteratively, element-wise, on the coefficients of the feature maps to optimise the CSC cost function. The performance of the proposed Iterative Cauchy Thresholding (ICT) algorithm in reconstructing natural images is compared against the common choice of 1\ell_1-norm optimised via soft and hard thresholding. ICT outperforms IHT and IST in most of these reconstruction experiments across various datasets, with an average PSNR of up to 11.30 and 7.04 above ISTA and IHT respectively.Comment: 19 pages, 9 figures, journal draf

    Representation Learning via Cauchy Convolutional Sparse Coding

    Get PDF
    In representation learning, Convolutional Sparse Coding (CSC) enables unsupervised learning of features by jointly optimising both an 2\ell_2-norm fidelity term and a sparsity enforcing penalty. This work investigates using a regularisation term derived from an assumed Cauchy prior for the coefficients of the feature maps of a CSC generative model. The sparsity penalty term resulting from this prior is solved via its proximal operator, which is then applied iteratively, element-wise, on the coefficients of the feature maps to optimise the CSC cost function. The performance of the proposed Iterative Cauchy Thresholding (ICT) algorithm in reconstructing natural images is compared against the common choice of 1\ell_1-norm optimised via soft and hard thresholding. ICT outperforms IHT and IST in most of these reconstruction experiments across various datasets, with an average PSNR of up to 11.30 and 7.04 above ISTA and IHT respectively.Comment: 19 pages, 9 figures, journal draf

    Rethinking Out-of-distribution (OOD) Detection: Masked Image Modeling is All You Need

    Full text link
    The core of out-of-distribution (OOD) detection is to learn the in-distribution (ID) representation, which is distinguishable from OOD samples. Previous work applied recognition-based methods to learn the ID features, which tend to learn shortcuts instead of comprehensive representations. In this work, we find surprisingly that simply using reconstruction-based methods could boost the performance of OOD detection significantly. We deeply explore the main contributors of OOD detection and find that reconstruction-based pretext tasks have the potential to provide a generally applicable and efficacious prior, which benefits the model in learning intrinsic data distributions of the ID dataset. Specifically, we take Masked Image Modeling as a pretext task for our OOD detection framework (MOOD). Without bells and whistles, MOOD outperforms previous SOTA of one-class OOD detection by 5.7%, multi-class OOD detection by 3.0%, and near-distribution OOD detection by 2.1%. It even defeats the 10-shot-per-class outlier exposure OOD detection, although we do not include any OOD samples for our detectionComment: This paper is accepted by CVPR2023 and our codes are released here: https://github.com/JulietLJY/MOO
    corecore