1,365,835 research outputs found
Hypernetwork functional image representation
Motivated by the human way of memorizing images we introduce their functional
representation, where an image is represented by a neural network. For this
purpose, we construct a hypernetwork which takes an image and returns weights
to the target network, which maps point from the plane (representing positions
of the pixel) into its corresponding color in the image. Since the obtained
representation is continuous, one can easily inspect the image at various
resolutions and perform on it arbitrary continuous operations. Moreover, by
inspecting interpolations we show that such representation has some properties
characteristic to generative models. To evaluate the proposed mechanism
experimentally, we apply it to image super-resolution problem. Despite using a
single model for various scaling factors, we obtained results comparable to
existing super-resolution methods
Sparse Image Representation with Epitomes
Sparse coding, which is the decomposition of a vector using only a few basis
elements, is widely used in machine learning and image processing. The basis
set, also called dictionary, is learned to adapt to specific data. This
approach has proven to be very effective in many image processing tasks.
Traditionally, the dictionary is an unstructured "flat" set of atoms. In this
paper, we study structured dictionaries which are obtained from an epitome, or
a set of epitomes. The epitome is itself a small image, and the atoms are all
the patches of a chosen size inside this image. This considerably reduces the
number of parameters to learn and provides sparse image decompositions with
shiftinvariance properties. We propose a new formulation and an algorithm for
learning the structured dictionaries associated with epitomes, and illustrate
their use in image denoising tasks.Comment: Computer Vision and Pattern Recognition, Colorado Springs : United
States (2011
Structured Knowledge Representation for Image Retrieval
We propose a structured approach to the problem of retrieval of images by
content and present a description logic that has been devised for the semantic
indexing and retrieval of images containing complex objects. As other
approaches do, we start from low-level features extracted with image analysis
to detect and characterize regions in an image. However, in contrast with
feature-based approaches, we provide a syntax to describe segmented regions as
basic objects and complex objects as compositions of basic ones. Then we
introduce a companion extensional semantics for defining reasoning services,
such as retrieval, classification, and subsumption. These services can be used
for both exact and approximate matching, using similarity measures. Using our
logical approach as a formal specification, we implemented a complete
client-server image retrieval system, which allows a user to pose both queries
by sketch and queries by example. A set of experiments has been carried out on
a testbed of images to assess the retrieval capabilities of the system in
comparison with expert users ranking. Results are presented adopting a
well-established measure of quality borrowed from textual information
retrieval
Anisotropic Mesh Adaptation for Image Representation
Triangular meshes have gained much interest in image representation and have
been widely used in image processing. This paper introduces a framework of
anisotropic mesh adaptation (AMA) methods to image representation and proposes
a GPRAMA method that is based on AMA and greedy-point removal (GPR) scheme.
Different than many other methods that triangulate sample points to form the
mesh, the AMA methods start directly with a triangular mesh and then adapt the
mesh based on a user-defined metric tensor to represent the image. The AMA
methods have clear mathematical framework and provides flexibility for both
image representation and image reconstruction. A mesh patching technique is
developed for the implementation of the GPRAMA method, which leads to an
improved version of the popular GPRFS-ED method. The GPRAMA method can achieve
better quality than the GPRFS-ED method but with lower computational cost.Comment: 25 pages, 15 figure
Object Level Deep Feature Pooling for Compact Image Representation
Convolutional Neural Network (CNN) features have been successfully employed
in recent works as an image descriptor for various vision tasks. But the
inability of the deep CNN features to exhibit invariance to geometric
transformations and object compositions poses a great challenge for image
search. In this work, we demonstrate the effectiveness of the objectness prior
over the deep CNN features of image regions for obtaining an invariant image
representation. The proposed approach represents the image as a vector of
pooled CNN features describing the underlying objects. This representation
provides robustness to spatial layout of the objects in the scene and achieves
invariance to general geometric transformations, such as translation, rotation
and scaling. The proposed approach also leads to a compact representation of
the scene, making each image occupy a smaller memory footprint. Experiments
show that the proposed representation achieves state of the art retrieval
results on a set of challenging benchmark image datasets, while maintaining a
compact representation.Comment: Deep Vision 201
- …