770 research outputs found
Higher-order neural network software for distortion invariant object recognition
The state-of-the-art in pattern recognition for such applications as automatic target recognition and industrial robotic vision relies on digital image processing. We present a higher-order neural network model and software which performs the complete feature extraction-pattern classification paradigm required for automatic pattern recognition. Using a third-order neural network, we demonstrate complete, 100 percent accurate invariance to distortions of scale, position, and in-plate rotation. In a higher-order neural network, feature extraction is built into the network, and does not have to be learned. Only the relatively simple classification step must be learned. This is key to achieving very rapid training. The training set is much smaller than with standard neural network software because the higher-order network only has to be shown one view of each object to be learned, not every possible view. The software and graphical user interface run on any Sun workstation. Results of the use of the neural software in autonomous robotic vision systems are presented. Such a system could have extensive application in robotic manufacturing
Advances in Hyperspectral Image Classification: Earth monitoring with statistical learning methods
Hyperspectral images show similar statistical properties to natural grayscale
or color photographic images. However, the classification of hyperspectral
images is more challenging because of the very high dimensionality of the
pixels and the small number of labeled examples typically available for
learning. These peculiarities lead to particular signal processing problems,
mainly characterized by indetermination and complex manifolds. The framework of
statistical learning has gained popularity in the last decade. New methods have
been presented to account for the spatial homogeneity of images, to include
user's interaction via active learning, to take advantage of the manifold
structure with semisupervised learning, to extract and encode invariances, or
to adapt classifiers and image representations to unseen yet similar scenes.
This tutuorial reviews the main advances for hyperspectral remote sensing image
classification through illustrative examples.Comment: IEEE Signal Processing Magazine, 201
Siamese Instance Search for Tracking
In this paper we present a tracker, which is radically different from
state-of-the-art trackers: we apply no model updating, no occlusion detection,
no combination of trackers, no geometric matching, and still deliver
state-of-the-art tracking performance, as demonstrated on the popular online
tracking benchmark (OTB) and six very challenging YouTube videos. The presented
tracker simply matches the initial patch of the target in the first frame with
candidates in a new frame and returns the most similar patch by a learned
matching function. The strength of the matching function comes from being
extensively trained generically, i.e., without any data of the target, using a
Siamese deep neural network, which we design for tracking. Once learned, the
matching function is used as is, without any adapting, to track previously
unseen targets. It turns out that the learned matching function is so powerful
that a simple tracker built upon it, coined Siamese INstance search Tracker,
SINT, which only uses the original observation of the target from the first
frame, suffices to reach state-of-the-art performance. Further, we show the
proposed tracker even allows for target re-identification after the target was
absent for a complete video shot.Comment: This paper is accepted to the IEEE Conference on Computer Vision and
Pattern Recognition, 201
Sign and Basis Invariant Networks for Spectral Graph Representation Learning
Many machine learning tasks involve processing eigenvectors derived from
data. Especially valuable are Laplacian eigenvectors, which capture useful
structural information about graphs and other geometric objects. However,
ambiguities arise when computing eigenvectors: for each eigenvector , the
sign flipped is also an eigenvector. More generally, higher dimensional
eigenspaces contain infinitely many choices of basis eigenvectors. These
ambiguities make it a challenge to process eigenvectors and eigenspaces in a
consistent way. In this work we introduce SignNet and BasisNet -- new neural
architectures that are invariant to all requisite symmetries and hence process
collections of eigenspaces in a principled manner. Our networks are universal,
i.e., they can approximate any continuous function of eigenvectors with the
proper invariances. They are also theoretically strong for graph representation
learning -- they can approximate any spectral graph convolution, can compute
spectral invariants that go beyond message passing neural networks, and can
provably simulate previously proposed graph positional encodings. Experiments
show the strength of our networks for molecular graph regression, learning
expressive graph representations, and learning implicit neural representations
on triangle meshes. Our code is available at
https://github.com/cptq/SignNet-BasisNet .Comment: 35 page
- …