30,756 research outputs found
Data Streams from the Low Frequency Instrument On-Board the Planck Satellite: Statistical Analysis and Compression Efficiency
The expected data rate produced by the Low Frequency Instrument (LFI) planned
to fly on the ESA Planck mission in 2007, is over a factor 8 larger than the
bandwidth allowed by the spacecraft transmission system to download the LFI
data. We discuss the application of lossless compression to Planck/LFI data
streams in order to reduce the overall data flow. We perform both theoretical
analysis and experimental tests using realistically simulated data streams in
order to fix the statistical properties of the signal and the maximal
compression rate allowed by several lossless compression algorithms. We studied
the influence of signal composition and of acquisition parameters on the
compression rate Cr and develop a semiempirical formalism to account for it.
The best performing compressor tested up to now is the arithmetic compression
of order 1, designed for optimizing the compression of white noise like
signals, which allows an overall compression rate = 2.65 +/- 0.02. We find
that such result is not improved by other lossless compressors, being the
signal almost white noise dominated. Lossless compression algorithms alone will
not solve the bandwidth problem but needs to be combined with other techniques.Comment: May 3, 2000 release, 61 pages, 6 figures coded as eps, 9 tables (4
included as eps), LaTeX 2.09 + assms4.sty, style file included, submitted for
the pubblication on PASP May 3, 200
Feasibility and performances of compressed-sensing and sparse map-making with Herschel/PACS data
The Herschel Space Observatory of ESA was launched in May 2009 and is in
operation since. From its distant orbit around L2 it needs to transmit a huge
quantity of information through a very limited bandwidth. This is especially
true for the PACS imaging camera which needs to compress its data far more than
what can be achieved with lossless compression. This is currently solved by
including lossy averaging and rounding steps on board. Recently, a new theory
called compressed-sensing emerged from the statistics community. This theory
makes use of the sparsity of natural (or astrophysical) images to optimize the
acquisition scheme of the data needed to estimate those images. Thus, it can
lead to high compression factors.
A previous article by Bobin et al. (2008) showed how the new theory could be
applied to simulated Herschel/PACS data to solve the compression requirement of
the instrument. In this article, we show that compressed-sensing theory can
indeed be successfully applied to actual Herschel/PACS data and give
significant improvements over the standard pipeline. In order to fully use the
redundancy present in the data, we perform full sky map estimation and
decompression at the same time, which cannot be done in most other compression
methods. We also demonstrate that the various artifacts affecting the data
(pink noise, glitches, whose behavior is a priori not well compatible with
compressed-sensing) can be handled as well in this new framework. Finally, we
make a comparison between the methods from the compressed-sensing scheme and
data acquired with the standard compression scheme. We discuss improvements
that can be made on ground for the creation of sky maps from the data.Comment: 11 pages, 6 figures, 5 tables, peer-reviewed articl
Designing a resource-efficient data structure for mobile data systems
Designing data structures for use in mobile devices requires attention on optimising data volumes with associated benefits for data transmission, storage space and battery use. For semi-structured data, tree summarisation techniques can be used to reduce the volume of structured elements while dictionary compression can efficiently deal with value-based predicates. This project seeks to investigate and evaluate an integration of the two approaches. The key strength of this technique is that both structural and value predicates could be resolved within one graph while further allowing for compression of the resulting data structure. As the current trend is towards the requirement for working with larger semi-structured data sets this work would allow for the utilisation of much larger data sets whilst reducing requirements on bandwidth and minimising the memory necessary both for the storage and querying of the data
Medical imaging analysis with artificial neural networks
Given that neural networks have been widely reported in the research community of medical imaging, we provide a focused literature survey on recent neural network developments in computer-aided diagnosis, medical image segmentation and edge detection towards visual content analysis, and medical image registration for its pre-processing and post-processing, with the aims of increasing awareness of how neural networks can be applied to these areas and to provide a foundation for further research and practical development. Representative techniques and algorithms are explained in detail to provide inspiring examples illustrating: (i) how a known neural network with fixed structure and training procedure could be applied to resolve a medical imaging problem; (ii) how medical images could be analysed, processed, and characterised by neural networks; and (iii) how neural networks could be expanded further to resolve problems relevant to medical imaging. In the concluding section, a highlight of comparisons among many neural network applications is included to provide a global view on computational intelligence with neural networks in medical imaging
Design of a digital compression technique for shuttle television
The determination of the performance and hardware complexity of data compression algorithms applicable to color television signals, were studied to assess the feasibility of digital compression techniques for shuttle communications applications. For return link communications, it is shown that a nonadaptive two dimensional DPCM technique compresses the bandwidth of field-sequential color TV to about 13 MBPS and requires less than 60 watts of secondary power. For forward link communications, a facsimile coding technique is recommended which provides high resolution slow scan television on a 144 KBPS channel. The onboard decoder requires about 19 watts of secondary power
Self-calibrating d-scan: measuring ultrashort laser pulses on-target using an arbitrary pulse compressor
In most applications of ultrashort pulse lasers, temporal compressors are
used to achieve a desired pulse duration in a target or sample, and precise
temporal characterization is important. The dispersion-scan (d-scan) pulse
characterization technique usually involves using glass wedges to impart
variable, well-defined amounts of dispersion to the pulses, while measuring the
spectrum of a nonlinear signal produced by those pulses. This works very well
for broadband few-cycle pulses, but longer, narrower bandwidth pulses are much
more difficult to measure this way. Here we demonstrate the concept of
self-calibrating d-scan, which extends the applicability of the d-scan
technique to pulses of arbitrary duration, enabling their complete measurement
without prior knowledge of the introduced dispersion. In particular, we show
that the pulse compressors already employed in chirped pulse amplification
(CPA) systems can be used to simultaneously compress and measure the temporal
profile of the output pulses on-target in a simple way, without the need of
additional diagnostics or calibrations, while at the same time calibrating the
often-unknown differential dispersion of the compressor itself. We demonstrate
the technique through simulations and experiments under known conditions.
Finally, we apply it to the measurement and compression of 27.5 fs pulses from
a CPA laser.Comment: 11 pages, 5 figures, Scientific Reports, in pres
Development of advanced digital techniques for data acquisition processing and communication Interim scientific report
Design, video data characteristics, error control, and compression algorithms for Mars television mapping missio
- …