6,350 research outputs found

    Human Motion Capture Data Tailored Transform Coding

    Full text link
    Human motion capture (mocap) is a widely used technique for digitalizing human movements. With growing usage, compressing mocap data has received increasing attention, since compact data size enables efficient storage and transmission. Our analysis shows that mocap data have some unique characteristics that distinguish themselves from images and videos. Therefore, directly borrowing image or video compression techniques, such as discrete cosine transform, does not work well. In this paper, we propose a novel mocap-tailored transform coding algorithm that takes advantage of these features. Our algorithm segments the input mocap sequences into clips, which are represented in 2D matrices. Then it computes a set of data-dependent orthogonal bases to transform the matrices to frequency domain, in which the transform coefficients have significantly less dependency. Finally, the compression is obtained by entropy coding of the quantized coefficients and the bases. Our method has low computational cost and can be easily extended to compress mocap databases. It also requires neither training nor complicated parameter setting. Experimental results demonstrate that the proposed scheme significantly outperforms state-of-the-art algorithms in terms of compression performance and speed

    Substructure Discovery Using Minimum Description Length and Background Knowledge

    Full text link
    The ability to identify interesting and repetitive substructures is an essential component to discovering knowledge in structural data. We describe a new version of our SUBDUE substructure discovery system based on the minimum description length principle. The SUBDUE system discovers substructures that compress the original data and represent structural concepts in the data. By replacing previously-discovered substructures in the data, multiple passes of SUBDUE produce a hierarchical description of the structural regularities in the data. SUBDUE uses a computationally-bounded inexact graph match that identifies similar, but not identical, instances of a substructure and finds an approximate measure of closeness of two substructures when under computational constraints. In addition to the minimum description length principle, other background knowledge can be used by SUBDUE to guide the search towards more appropriate substructures. Experiments in a variety of domains demonstrate SUBDUE's ability to find substructures capable of compressing the original data and to discover structural concepts important to the domain. Description of Online Appendix: This is a compressed tar file containing the SUBDUE discovery system, written in C. The program accepts as input databases represented in graph form, and will output discovered substructures with their corresponding value.Comment: See http://www.jair.org/ for an online appendix and other files accompanying this articl

    Compression of Deep Neural Networks on the Fly

    Full text link
    Thanks to their state-of-the-art performance, deep neural networks are increasingly used for object recognition. To achieve these results, they use millions of parameters to be trained. However, when targeting embedded applications the size of these models becomes problematic. As a consequence, their usage on smartphones or other resource limited devices is prohibited. In this paper we introduce a novel compression method for deep neural networks that is performed during the learning phase. It consists in adding an extra regularization term to the cost function of fully-connected layers. We combine this method with Product Quantization (PQ) of the trained weights for higher savings in storage consumption. We evaluate our method on two data sets (MNIST and CIFAR10), on which we achieve significantly larger compression rates than state-of-the-art methods

    Landsat Satellite Image Segmentation Using the Fuzzy ARTMAP Neural Network

    Full text link
    This application illustrates how the fuzzy ARTMAP neural network can be used to monitor environmental changes. A benchmark problem seeks to classify regions of a Landsat image into six soil and crop classes based on images from four spectral sensors. Simulations show that fuzzy ARTMAP outperforms fourteen other neural network and machine learning algorithms. Only the k-Nearest-Neighbor algorithm shows better performance (91% vs. 89%) but without any code compression, while fuzzy ARTMAP achieves a code compression ratio of 6:1. Even with a code compression ratio of 50:1 fuzzy ARTMAP still maintains good performance (83%). This example shows how fuzzy ARTMAP can combine accuracy and code compression in real-world applications.Office of Naval Research (N00014-92-J-401J, N00014-91-J-4100, N00014-92-J-4015); National Science Foundation (IRI 90-00530

    Landsat Satellite Image Segmentation Using the Fuzzy ARTMAP Neural Network

    Full text link
    This application illustrates how the fuzzy ARTMAP neural network can be used to monitor environmental changes. A benchmark problem seeks to classify regions of a Landsat image into six soil and crop classes based on images from four spectral sensors. Simulations show that fuzzy ARTMAP outperforms fourteen other neural network and machine learning algorithms. Only the k-Nearest-Neighbor algorithm shows better performance (91% vs. 89%) but without any code compression, while fuzzy ARTMAP achieves a code compression ratio of 6:1. Even with a code compression ratio of 50:1 fuzzy ARTMAP still maintains good performance (83%). This example shows how fuzzy ARTMAP can combine accuracy and code compression in real-world applications.Office of Naval Research (N00014-92-J-401J, N00014-91-J-4100, N00014-92-J-4015); National Science Foundation (IRI 90-00530

    The SP theory of intelligence: benefits and applications

    Full text link
    This article describes existing and expected benefits of the "SP theory of intelligence", and some potential applications. The theory aims to simplify and integrate ideas across artificial intelligence, mainstream computing, and human perception and cognition, with information compression as a unifying theme. It combines conceptual simplicity with descriptive and explanatory power across several areas of computing and cognition. In the "SP machine" -- an expression of the SP theory which is currently realized in the form of a computer model -- there is potential for an overall simplification of computing systems, including software. The SP theory promises deeper insights and better solutions in several areas of application including, most notably, unsupervised learning, natural language processing, autonomous robots, computer vision, intelligent databases, software engineering, information compression, medical diagnosis and big data. There is also potential in areas such as the semantic web, bioinformatics, structuring of documents, the detection of computer viruses, data fusion, new kinds of computer, and the development of scientific theories. The theory promises seamless integration of structures and functions within and between different areas of application. The potential value, worldwide, of these benefits and applications is at least $190 billion each year. Further development would be facilitated by the creation of a high-parallel, open-source version of the SP machine, available to researchers everywhere.Comment: arXiv admin note: substantial text overlap with arXiv:1212.022

    Fine-Pruning: Joint Fine-Tuning and Compression of a Convolutional Network with Bayesian Optimization

    Full text link
    When approaching a novel visual recognition problem in a specialized image domain, a common strategy is to start with a pre-trained deep neural network and fine-tune it to the specialized domain. If the target domain covers a smaller visual space than the source domain used for pre-training (e.g. ImageNet), the fine-tuned network is likely to be over-parameterized. However, applying network pruning as a post-processing step to reduce the memory requirements has drawbacks: fine-tuning and pruning are performed independently; pruning parameters are set once and cannot adapt over time; and the highly parameterized nature of state-of-the-art pruning methods make it prohibitive to manually search the pruning parameter space for deep networks, leading to coarse approximations. We propose a principled method for jointly fine-tuning and compressing a pre-trained convolutional network that overcomes these limitations. Experiments on two specialized image domains (remote sensing images and describable textures) demonstrate the validity of the proposed approach.Comment: BMVC 2017 ora
    • …
    corecore