30,883 research outputs found
A reduced algorithm from Faugeras-Bethods theorem in labeling problems
This paper illustrates a new approach to labeling ("object classification") problems, and it targets the simplification of a (computationally) complex algorithm based on Faugeras and Berthod's theorem.
Parallel Image Processing Using a Pure Topological Framework
Image processing is a fundamental operation
in many real time applications, where lots of parallelism
can be extracted. Segmenting the image into different
connected components is the most known operations, but
there are many others like extracting the region adjacency
graph (RAG) of these regions, or searching for features
points, being invariant to rotations, scales, brilliant
changes, etc. Most of these algorithms part from the basis
of Tracing-type approaches or scan/raster methods. This
fact necessarily implies a data dependence between the
processing of one pixel and the previous one, which
prevents using a pure parallel approach. In terms of time
complexity, this means that linear order O(N) (N being the
number of pixels) cannot be cut down. In this paper, we
describe a novel approach based on the building of a pure
Topological framework, which allows to implement fully
parallel algorithms. Concerning topological analysis, a first
stage is computed in parallel for every pixel, thus
conveying the local neighboring conditions. Then, they are
extended in a second parallel stage to the necessary global
relations (e.g. to join all the pixels of a connected
component). This combinatorial optimization process can
be seen as the compression of the whole image to just one
pixel. Using this final representation, every region can be
related with the rest, which yields to pure topological
construction of other image operations. Besides, complex
data structures can be avoided: all the processing can be
done using matrixes (with the same indexation as the
original image) and element-wise operations. The time
complexity order of our topological approach for a mĂ—n
pixel image is near O(log(m+n)), under the assumption that
a processing element exists for each pixel. Results for a
multicore processor show very good scalability until the
memory bandwidth bottleneck is reached, both for bigger
images and for much optimized implementations. The
inherent parallelism of our approach points to the
direction that even better results will be obtained in other
less classical computing architectures.1Ministerio de EconomĂa y Competitividad (España) TEC2012-37868-C04-02AEI/FEDER (UE) MTM2016-81030-PVPPI of the University of Sevill
Image mining: trends and developments
[Abstract]: Advances in image acquisition and storage technology have led to tremendous growth in very large and detailed image databases. These images, if analyzed, can reveal useful information to the human users. Image mining deals with the extraction of implicit knowledge, image data relationship, or other patterns not explicitly stored in the images. Image mining is more than just an extension of data mining to image domain. It is an interdisciplinary endeavor that draws upon expertise in computer vision, image processing, image retrieval, data mining, machine learning, database, and artificial intelligence. In this paper, we will examine the research issues in image mining, current developments in image mining, particularly, image mining frameworks, state-of-the-art techniques and systems. We will also identify some future research directions for image mining
Recommended from our members
Recognition by directed attention to recursively partitioned images
A learning/recognition model (and instantiating program) is described which recursively combines the learning paradigms of conceptual clustering (Michalski, 1980) and learning-from-examples to resolve the ambiguities of real-world recognition. The model is based on neuropsychological and psychological evidence that the visual system is analytic, hierarchical, and composed of a parallel/serial dichotomy (many, see conclusions by Crick, 1984). Emulating the experimental evidence, parallel processes in the model decompose the image into components and cluster the constituents in much the same way as the image processing technique known as moment analysis (Alt, 1962). Serial, attentive mechanisms then reassemble the decompositions by investigating spatial relationships between components. The use of attentive mechanisms extends the moment analysis technique to handle alterations in structure and solves the contention problem created by combining the two learning paradigms. The contention results from a disagreement between the teacher and the model on what constitutes the salient features at the highest level of the symbol. There are four cases ZBT must handle, two of which result from the disagreement with the teacher. The parallel/serial dichotomy represents a vertical/horizontal tradeoff between the invariant and variant features of a domain. The resultant learned hierarchy allows ZBT to recognize structural differences while avoiding problems of exponential growth
Markov Random Field Models: A Bayesian Approach to Computer Vision Problems
The object of our study is the Bayesian approach in solving computer vision problems. We examine in particular: (i) applications of Markov random field (MRF) models to modeling spatial images; (ii) MRF based statistical methods for image restoration, segmentation, texture modeling and integration of different visual cues
- …