694 research outputs found
Complex data processing: fast wavelet analysis on the sphere
In the general context of complex data processing, this paper reviews a
recent practical approach to the continuous wavelet formalism on the sphere.
This formalism notably yields a correspondence principle which relates wavelets
on the plane and on the sphere. Two fast algorithms are also presented for the
analysis of signals on the sphere with steerable wavelets.Comment: 20 pages, 5 figures, JFAA style, paper invited to J. Fourier Anal.
and Appli
A survey of exemplar-based texture synthesis
Exemplar-based texture synthesis is the process of generating, from an input
sample, new texture images of arbitrary size and which are perceptually
equivalent to the sample. The two main approaches are statistics-based methods
and patch re-arrangement methods. In the first class, a texture is
characterized by a statistical signature; then, a random sampling conditioned
to this signature produces genuinely different texture images. The second class
boils down to a clever "copy-paste" procedure, which stitches together large
regions of the sample. Hybrid methods try to combine ideas from both approaches
to avoid their hurdles. The recent approaches using convolutional neural
networks fit to this classification, some being statistical and others
performing patch re-arrangement in the feature space. They produce impressive
synthesis on various kinds of textures. Nevertheless, we found that most real
textures are organized at multiple scales, with global structures revealed at
coarse scales and highly varying details at finer ones. Thus, when confronted
with large natural images of textures the results of state-of-the-art methods
degrade rapidly, and the problem of modeling them remains wide open.Comment: v2: Added comments and typos fixes. New section added to describe
FRAME. New method presented: CNNMR
Recommended from our members
Coherence of Multiscale Features for Enhancement of Digital Mammograms
Mammograms depict most of the significant changes in breast disease. The primary radiographic signs of cancer are related to tumor mass, density, size, borders, and shape, and local distribution of calcifications. We show that each of these features can be well described by coherence and orientation measures and provide visual cues for radiologists to identify possible lesions more easily without increasing false positives. In this paper, an artifact-free enhancement algorithm based on overcomplete multiscale representations is presented. First, an image was decomposed using a fast wavelet transform algorithm. At each level of analysis, energy and phase information are computed via a set of separable steerable filters. Then, a measure of coherence within each level was obtained by weighting an energy measure with the ratio of projections of local energy within a specified window. Each projection was computed onto the central point of a window with respect to the total energy within that window. Finally, a nonlinear operation, integrating coherence and orientation information, was applied to modify transform coefficients within distinct levels of analysis. These modified coefficients were then reconstructed, via an inverse fast wavelet transform, resulting in an improved visualization of significant mammographic features. The novelty of this algorithm lies in the detection of directional multiscale features and the removal of aliased perturbations
A Panorama on Multiscale Geometric Representations, Intertwining Spatial, Directional and Frequency Selectivity
The richness of natural images makes the quest for optimal representations in
image processing and computer vision challenging. The latter observation has
not prevented the design of image representations, which trade off between
efficiency and complexity, while achieving accurate rendering of smooth regions
as well as reproducing faithful contours and textures. The most recent ones,
proposed in the past decade, share an hybrid heritage highlighting the
multiscale and oriented nature of edges and patterns in images. This paper
presents a panorama of the aforementioned literature on decompositions in
multiscale, multi-orientation bases or dictionaries. They typically exhibit
redundancy to improve sparsity in the transformed domain and sometimes its
invariance with respect to simple geometric deformations (translation,
rotation). Oriented multiscale dictionaries extend traditional wavelet
processing and may offer rotation invariance. Highly redundant dictionaries
require specific algorithms to simplify the search for an efficient (sparse)
representation. We also discuss the extension of multiscale geometric
decompositions to non-Euclidean domains such as the sphere or arbitrary meshed
surfaces. The etymology of panorama suggests an overview, based on a choice of
partially overlapping "pictures". We hope that this paper will contribute to
the appreciation and apprehension of a stream of current research directions in
image understanding.Comment: 65 pages, 33 figures, 303 reference
Deformable kernels for early vision
Early vision algorithms often have a first stage of linear-filtering that `extracts' from the image information at multiple scales of resolution and multiple orientations. A common difficulty in the design and implementation of such schemes is that one feels compelled to discretize coarsely the space of scales and orientations in order to reduce computation and storage costs. A technique is presented that allows: 1) computing the best approximation of a given family using linear combinations of a small number of `basis' functions; and 2) describing all finite-dimensional families, i.e., the families of filters for which a finite dimensional representation is possible with no error. The technique is based on singular value decomposition and may be applied to generating filters in arbitrary dimensions and subject to arbitrary deformations. The relevant functional analysis results are reviewed and precise conditions for the decomposition to be feasible are stated. Experimental results are presented that demonstrate the applicability of the technique to generating multiorientation multi-scale 2D edge-detection kernels. The implementation issues are also discussed
Deformable kernels for early vision
Early vision algorithms often have a first stage of linear filtering
that 'extracts' from the image information at multiple
scales of resolution and multiple orientations. A common
difficulty in the design and implementation of such
schemes is that one feels compelled to discretize coarsely
the space of scales and orientations in order to reduce computation and storage costs. This discretization produces
anisotropies due to a loss of traslation-, rotation- scaling- invariance that makes early vision algorithms less precise and
more difficult to design. This need not be so: one can compute
and store efficiently the response of families of linear
filters defined on a continuum of orientations and scales. A
technique is presented that allows (1) to compute the best approximation of a given family using linear combinations of
a small number of 'basis' functions; (2) to describe all finite-dimensional families, i.e. the families of filters for which a
finite dimensional representation is possible with no error.
The technique is general and can be applied to generating
filters in arbitrary dimensions. Experimental results are presented that demonstrate the applicability of the technique to
generating multi-orientation multi-scale 20 edge-detection
kernels. The implementation issues are also discussed
Dynamic Steerable Blocks in Deep Residual Networks
Filters in convolutional networks are typically parameterized in a pixel
basis, that does not take prior knowledge about the visual world into account.
We investigate the generalized notion of frames designed with image properties
in mind, as alternatives to this parametrization. We show that frame-based
ResNets and Densenets can improve performance on Cifar-10+ consistently, while
having additional pleasant properties like steerability. By exploiting these
transformation properties explicitly, we arrive at dynamic steerable blocks.
They are an extension of residual blocks, that are able to seamlessly transform
filters under pre-defined transformations, conditioned on the input at training
and inference time. Dynamic steerable blocks learn the degree of invariance
from data and locally adapt filters, allowing them to apply a different
geometrical variant of the same filter to each location of the feature map.
When evaluated on the Berkeley Segmentation contour detection dataset, our
approach outperforms all competing approaches that do not utilize pre-training.
Our results highlight the benefits of image-based regularization to deep
networks
A matlab toolbox for image fusion (MATIFUS).
The MATIFUS toolbox is presented. It is a collection of functions and furnished with a graphical user interface that supports a range of image fusion operations. Almost all of the toolbox functions are written in the MATLAB language. Implementations of multiresolution schemes are used that are either publicly available or can be purchased as licensed software. MATIFUS can be downloaded from a website and is available under the conditions of an agreement with the Dutch Technology Foundation ST
- …