6,552 research outputs found

    Prioritized Data Compression using Wavelets

    Full text link
    The volume of data and the velocity with which it is being generated by com- putational experiments on high performance computing (HPC) systems is quickly outpacing our ability to effectively store this information in its full fidelity. There- fore, it is critically important to identify and study compression methodologies that retain as much information as possible, particularly in the most salient regions of the simulation space. In this paper, we cast this in terms of a general decision-theoretic problem and discuss a wavelet-based compression strategy for its solution. We pro- vide a heuristic argument as justification and illustrate our methodology on several examples. Finally, we will discuss how our proposed methodology may be utilized in an HPC environment on large-scale computational experiments

    The Data Big Bang and the Expanding Digital Universe: High-Dimensional, Complex and Massive Data Sets in an Inflationary Epoch

    Get PDF
    Recent and forthcoming advances in instrumentation, and giant new surveys, are creating astronomical data sets that are not amenable to the methods of analysis familiar to astronomers. Traditional methods are often inadequate not merely because of the size in bytes of the data sets, but also because of the complexity of modern data sets. Mathematical limitations of familiar algorithms and techniques in dealing with such data sets create a critical need for new paradigms for the representation, analysis and scientific visualization (as opposed to illustrative visualization) of heterogeneous, multiresolution data across application domains. Some of the problems presented by the new data sets have been addressed by other disciplines such as applied mathematics, statistics and machine learning and have been utilized by other sciences such as space-based geosciences. Unfortunately, valuable results pertaining to these problems are mostly to be found only in publications outside of astronomy. Here we offer brief overviews of a number of concepts, techniques and developments, some "old" and some new. These are generally unknown to most of the astronomical community, but are vital to the analysis and visualization of complex datasets and images. In order for astronomers to take advantage of the richness and complexity of the new era of data, and to be able to identify, adopt, and apply new solutions, the astronomical community needs a certain degree of awareness and understanding of the new concepts. One of the goals of this paper is to help bridge the gap between applied mathematics, artificial intelligence and computer science on the one side and astronomy on the other.Comment: 24 pages, 8 Figures, 1 Table. Accepted for publication: "Advances in Astronomy, special issue "Robotic Astronomy

    Review of analytical instruments for EEG analysis

    Full text link
    Since it was first used in 1926, EEG has been one of the most useful instruments of neuroscience. In order to start using EEG data we need not only EEG apparatus, but also some analytical tools and skills to understand what our data mean. This article describes several classical analytical tools and also new one which appeared only several years ago. We hope it will be useful for those researchers who have only started working in the field of cognitive EEG

    MADNESS: A Multiresolution, Adaptive Numerical Environment for Scientific Simulation

    Full text link
    MADNESS (multiresolution adaptive numerical environment for scientific simulation) is a high-level software environment for solving integral and differential equations in many dimensions that uses adaptive and fast harmonic analysis methods with guaranteed precision based on multiresolution analysis and separated representations. Underpinning the numerical capabilities is a powerful petascale parallel programming environment that aims to increase both programmer productivity and code scalability. This paper describes the features and capabilities of MADNESS and briefly discusses some current applications in chemistry and several areas of physics

    Improved Stroke Detection at Early Stages Using Haar Wavelets and Laplacian Pyramid

    Get PDF
    Stroke merupakan pembunuh nomor tiga di dunia, namun hanya sedikit metode tentang deteksi dini. Oleh karena itu dibutuhkan metode untuk mendeteksi hal tersebut. Penelitian ini mengusulkan sebuah metode gabungan untuk mendeteksi dua jenis stroke secara simultan. Haar wavelets untuk mendeteksi stroke hemoragik dan Laplacian pyramid untuk mendeteksi stroke iskemik. Tahapan dalam penelitian ini terdiri dari pra proses tahap 1 dan 2, Haar wavelets, Laplacian pyramid, dan perbaikan kualitas citra. Pra proses adalah menghilangkan bagian tulang tengkorak, reduksi derau, perbaikan kontras, dan menghilangkan bagian selain citra otak. Kemudian dilakukan perbaikan citra. Selanjutnya Haar wavelet digunakan untuk ekstraksi daerah hemoragik sedangkan Laplacian pyramid untuk ekstraksi daerah iskemik. Tahapan terakhir adalah menghitung fitur Grey Level Cooccurrence Matrix (GLCM) sebagai fitur untuk proses klasifikasi. Hasil visualisasi diproses lanjut untuk ekstrasi fitur menggunakan GLCM dengan 12 fitur dan kemudian GLCM dengan 4 fitur. Untuk proses klasifikasi digunakan SVM dan KNN, sedangkan pengukuran performa menggunakan akurasi. Jumlah data hemoragik dan iskemik adalah 45 citra yang dibagi menjadi 2 bagian, 28 citra untuk pengujian dan 17 citra untuk pelatihan. Hasil akhir menunjukkan akurasi tertinggi yang dicapai menggunakan SVM adalah 82% dan KNN adalah 88%
    corecore