347 research outputs found
Scanning and Sequential Decision Making for Multi-Dimensional Data - Part I: the Noiseless Case
We investigate the problem of scanning and prediction ("scandiction", for
short) of multidimensional data arrays. This problem arises in several aspects
of image and video processing, such as predictive coding, for example, where an
image is compressed by coding the error sequence resulting from scandicting it.
Thus, it is natural to ask what is the optimal method to scan and predict a
given image, what is the resulting minimum prediction loss, and whether there
exist specific scandiction schemes which are universal in some sense.
Specifically, we investigate the following problems: First, modeling the data
array as a random field, we wish to examine whether there exists a scandiction
scheme which is independent of the field's distribution, yet asymptotically
achieves the same performance as if this distribution was known. This question
is answered in the affirmative for the set of all spatially stationary random
fields and under mild conditions on the loss function. We then discuss the
scenario where a non-optimal scanning order is used, yet accompanied by an
optimal predictor, and derive bounds on the excess loss compared to optimal
scanning and prediction.
This paper is the first part of a two-part paper on sequential decision
making for multi-dimensional data. It deals with clean, noiseless data arrays.
The second part deals with noisy data arrays, namely, with the case where the
decision maker observes only a noisy version of the data, yet it is judged with
respect to the original, clean data.Comment: 46 pages, 2 figures. Revised version: title changed, section 1
revised, section 3.1 added, a few minor/technical corrections mad
Discrete denoising of heterogenous two-dimensional data
We consider discrete denoising of two-dimensional data with characteristics
that may be varying abruptly between regions.
Using a quadtree decomposition technique and space-filling curves, we extend
the recently developed S-DUDE (Shifting Discrete Universal DEnoiser), which was
tailored to one-dimensional data, to the two-dimensional case. Our scheme
competes with a genie that has access, in addition to the noisy data, also to
the underlying noiseless data, and can employ different two-dimensional
sliding window denoisers along distinct regions obtained by a quadtree
decomposition with leaves, in a way that minimizes the overall loss. We
show that, regardless of what the underlying noiseless data may be, the
two-dimensional S-DUDE performs essentially as well as this genie, provided
that the number of distinct regions satisfies , where is the total
size of the data. The resulting algorithm complexity is still linear in both
and , as in the one-dimensional case. Our experimental results show that
the two-dimensional S-DUDE can be effective when the characteristics of the
underlying clean image vary across different regions in the data.Comment: 16 pages, submitted to IEEE Transactions on Information Theor
A Study of Energy and Locality Effects using Space-filling Curves
The cost of energy is becoming an increasingly important driver for the
operating cost of HPC systems, adding yet another facet to the challenge of
producing efficient code. In this paper, we investigate the energy implications
of trading computation for locality using Hilbert and Morton space-filling
curves with dense matrix-matrix multiplication. The advantage of these curves
is that they exhibit an inherent tiling effect without requiring specific
architecture tuning. By accessing the matrices in the order determined by the
space-filling curves, we can trade computation for locality. The index
computation overhead of the Morton curve is found to be balanced against its
locality and energy efficiency, while the overhead of the Hilbert curve
outweighs its improvements on our test system.Comment: Proceedings of the 2014 IEEE International Parallel & Distributed
Processing Symposium Workshops (IPDPSW
Harmonious Hilbert curves and other extradimensional space-filling curves
This paper introduces a new way of generalizing Hilbert's two-dimensional
space-filling curve to arbitrary dimensions. The new curves, called harmonious
Hilbert curves, have the unique property that for any d' < d, the d-dimensional
curve is compatible with the d'-dimensional curve with respect to the order in
which the curves visit the points of any d'-dimensional axis-parallel space
that contains the origin. Similar generalizations to arbitrary dimensions are
described for several variants of Peano's curve (the original Peano curve, the
coil curve, the half-coil curve, and the Meurthe curve). The d-dimensional
harmonious Hilbert curves and the Meurthe curves have neutral orientation: as
compared to the curve as a whole, arbitrary pieces of the curve have each of d!
possible rotations with equal probability. Thus one could say these curves are
`statistically invariant' under rotation---unlike the Peano curves, the coil
curves, the half-coil curves, and the familiar generalization of Hilbert curves
by Butz and Moore.
In addition, prompted by an application in the construction of R-trees, this
paper shows how to construct a 2d-dimensional generalized Hilbert or Peano
curve that traverses the points of a certain d-dimensional diagonally placed
subspace in the order of a given d-dimensional generalized Hilbert or Peano
curve.
Pseudocode is provided for comparison operators based on the curves presented
in this paper.Comment: 40 pages, 10 figures, pseudocode include
Adaptive pattern recognition by mini-max neural networks as a part of an intelligent processor
In this decade and progressing into 21st Century, NASA will have missions including Space Station and the Earth related Planet Sciences. To support these missions, a high degree of sophistication in machine automation and an increasing amount of data processing throughput rate are necessary. Meeting these challenges requires intelligent machines, designed to support the necessary automations in a remote space and hazardous environment. There are two approaches to designing these intelligent machines. One of these is the knowledge-based expert system approach, namely AI. The other is a non-rule approach based on parallel and distributed computing for adaptive fault-tolerances, namely Neural or Natural Intelligence (NI). The union of AI and NI is the solution to the problem stated above. The NI segment of this unit extracts features automatically by applying Cauchy simulated annealing to a mini-max cost energy function. The feature discovered by NI can then be passed to the AI system for future processing, and vice versa. This passing increases reliability, for AI can follow the NI formulated algorithm exactly, and can provide the context knowledge base as the constraints of neurocomputing. The mini-max cost function that solves the unknown feature can furthermore give us a top-down architectural design of neural networks by means of Taylor series expansion of the cost function. A typical mini-max cost function consists of the sample variance of each class in the numerator, and separation of the center of each class in the denominator. Thus, when the total cost energy is minimized, the conflicting goals of intraclass clustering and interclass segregation are achieved simultaneously
Joint denoising and distortion correction of atomic scale scanning transmission electron microscopy images
Nowadays, modern electron microscopes deliver images at atomic scale. The
precise atomic structure encodes information about material properties. Thus,
an important ingredient in the image analysis is to locate the centers of the
atoms shown in micrographs as precisely as possible. Here, we consider scanning
transmission electron microscopy (STEM), which acquires data in a rastering
pattern, pixel by pixel. Due to this rastering combined with the magnification
to atomic scale, movements of the specimen even at the nanometer scale lead to
random image distortions that make precise atom localization difficult. Given a
series of STEM images, we derive a Bayesian method that jointly estimates the
distortion in each image and reconstructs the underlying atomic grid of the
material by fitting the atom bumps with suitable bump functions. The resulting
highly non-convex minimization problems are solved numerically with a trust
region approach. Well-posedness of the reconstruction method and the model
behavior for faster and faster rastering are investigated using variational
techniques. The performance of the method is finally evaluated on both
synthetic and real experimental data
Neural Space-filling Curves
We present Neural Space-filling Curves (SFCs), a data-driven approach to
infer a context-based scan order for a set of images. Linear ordering of pixels
forms the basis for many applications such as video scrambling, compression,
and auto-regressive models that are used in generative modeling for images.
Existing algorithms resort to a fixed scanning algorithm such as Raster scan or
Hilbert scan. Instead, our work learns a spatially coherent linear ordering of
pixels from the dataset of images using a graph-based neural network. The
resulting Neural SFC is optimized for an objective suitable for the downstream
task when the image is traversed along with the scan line order. We show the
advantage of using Neural SFCs in downstream applications such as image
compression. Code and additional results will be made available at
https://hywang66.github.io/publication/neuralsfc
- …