12 research outputs found
Artificial Neural Networks as Decision-Makers for Stereo Matching
This paper investigates the use of artificial neural networks to help making a decision on matching of stereo images. An image matching technique based on extracting features from segmented regions is adopted in this work, and a neural network framework is applied for region matching of stereo photographs. Two types of neural networks are used, the radial basis network, (RB) for learning clustering, and the back propagation (BP) network for learning image matching. The (RB) neural network is to cluster the regions according to the locations of their centered points. For each region, the BP network uses differential features as input training data. While training and testing the system, multiple features are extracted and used for enhancing the accuracy of the matching process. Features include (compactness, Euler number, and invariant moments) for each region. Results obtained from the neural networks (namely; clustering and initial matching array) are used to select the best matching pair. Results are showing a good matching accuracy
Representations for Cognitive Vision : a Review of Appearance-Based, Spatio-Temporal, and Graph-Based Approaches
The emerging discipline of cognitive vision requires a proper representation of visual information including spatial and temporal relationships, scenes, events, semantics and context. This review article summarizes existing representational schemes in computer vision which might be useful for cognitive vision, a and discusses promising future research directions. The various approaches are categorized according to appearance-based, spatio-temporal, and graph-based representations for cognitive vision. While the representation of objects has been covered extensively in computer vision research, both from a reconstruction as well as from a recognition point of view, cognitive vision will also require new ideas how to represent scenes. We introduce new concepts for scene representations and discuss how these might be efficiently implemented in future cognitive vision systems
Automatic Positional Accuracy Assessment of Imagery Segmentation Processes: A Case Study
There are many studies related to Imagery Segmentation (IS) in the field of Geographic Information (GI). However, none of them address the assessment of IS results from a positional perspective. In a field in which the positional aspect is critical, it seems reasonable to think that the quality associated with this aspect must be controlled. This paper presents an automatic positional accuracy assessment (PAA) method for assessing this quality component of the regions obtained by means of the application of a textural segmentation algorithm to a Very High Resolution (VHR) aerial image. This method is based on the comparison between the ideal segmentation and the computed segmentation by counting their differences. Therefore, it has the same conceptual principles as the automatic procedures used in the evaluation of the GI's positional accuracy. As in any PAA method, there are two key aspects related to the sample that were addressed: (i) its size-specifically, its influence on the uncertainty of the estimated accuracy values-and (ii) its categorization. Although the results obtained must be taken with caution, they made it clear that automatic PAA procedures, which are mainly applied to carry out the positional quality assessment of cartography, are valid for assessing the positional accuracy reached using other types of processes. Such is the case of the IS process presented in this study
Representative discovery of structure cues for weakly-supervised image segmentation
National Research Foundation (NRF) Singapore under International Research Centre @ Singapore Funding Initiativ
Representative discovery of structure cues for weakly-supervised image segmentation
National Research Foundation (NRF) Singapore under International Research Centre @ Singapore Funding Initiativ
Recommended from our members
A Stochastic Grammar of Images
This exploratory paper quests for a stochastic and context sensitive grammar of images. The grammar should achieve the following four objectives and thus serves as a unified framework of representation, learning, and recognition for a large number of object categories. (i) The grammar represents both the hierarchical decompositions from scenes, to objects, parts, primitives and pixels by terminal and non-terminal nodes and the contexts for spatial and functional relations by horizontal links between the nodes. It formulates each object category as the set of all possible valid configurations produced by the grammar. (ii) The grammar is embodied in a simple And-Or graph representation where each Or-node points to alternative sub-configurations and an And-node is decomposed into a number of components. This representation supports recursive top-down/bottom-up procedures for image parsing under the Bayesian framework and make it convenient to scale up in complexity. Given an input image, the image parsing task constructs a most probable parse graph on-the-fly as the output interpretation and this parse graph is a subgraph of the And-Or graph after making choice on the Or-nodes. (iii) A probabilistic model is defined on this And-Or graph representation to account for the natural occurrence frequency of objects and parts as well as their relations. This model is learned from a relatively small training set per category and then sampled to synthesize a large number of configurations to cover novel object instances in the test set. This generalization capability is mostly missing in discriminative machine learning methods and can largely improve recognition performance in experiments. (iv) To fill the well-known semantic gap between symbols and raw signals, the grammar includes a series of visual dictionaries and organizes them through graph composition. At the bottom-level the dictionary is a set of image primitives each having a number of anchor points with open bonds to link with other primitives. These primitives can be combined to form larger and larger graph structures for parts and objects. The ambiguities in inferring local primitives shall be resolved through top-down computation using larger structures. Finally these primitives forms a primal sketch representation which will generate the input image with every pixels explained. The proposal grammar integrates three prominent representations in the literature: stochastic grammars for composition, Markov (or graphical) models for contexts, and sparse coding with primitives (wavelets). It also combines the structure-based and appearance based methods in the vision literature. Finally the paper presents three case studies to illustrate the proposed grammar.Mathematic