334 research outputs found

    Geometric Modeling and Recognition of Elongated Regions in Images.

    Get PDF
    The goal of this research is the recovery of elongated shapes from patterns of local features extracted from images. A generic geometric model-based approach is developed based on general concepts of 2-d form and structure. This is an intermediate-level analysis that is computed from groupings and decompositions of related low-level features. Axial representations are used to describe the shapes of image objects having the property of elongatedness. Curve-fitting is shown to compute axial sequences of the points in an elongated cluster. Script-clustering is performed about a parametric smooth curve to extract elongated partitions of the data incorporating constraints of point connectivity, curve alignment, and strip boundedness. A thresholded version of the Gabriel Graph (GG) is shown to offer most of the information needed from the Minimum Spanning Tree (MST) and Delauney Triangulation (DT), while being easily computable from finite neighborhood operations. An iterative curve-fitting method, that is placed in the general framework of Random Sample Consensus (RANSAC) model-fitting, is used to extract maximal partitions. The method is developed for general parametric curve-fitting over discrete point patterns. A complete structural analysis is presented for the recovery of elongated regions from multispectral classification. A region analysis is shown to be superior to an edge-based analysis in the early stages of recognition. First, the curve-fitting method is used to recover the linear components of complex object regions. The rough locations to start and end a region delineation are then detected by decomposing extracted linear shape clusters with a circular operator. Experimental results are shown for a variety of images, with the main result being an analysis of a high-resolution aerial image of a suburban road network. Analyses of printed circuit board patterns and a LANDSAT river image are also given. The generality of the curve-fitting approach is shown by these results and by its possible applications to other described image analysis problems

    Quantification of Order in Point Patterns

    Get PDF
    Pattern attributes are important in many disciplines, e.g. developmental biology, but there are few objective measures of them. Here we concentrate on the attribute of order in point patterns and its objective measurement. We examine perception of order and develop analysis algorithms that quantify the attribute in accordance with perception of it. Based on pairwise ranking of point patterns by degree of order, we show that judgements are highly consistent across individuals and that the perceptual dimension has an interval scale structure, spanning roughly 10 just-noticeable differences (jnds) between disorder and order. We designed a geometric algorithm that estimates order to an accuracy of half a jnd by quantifying the variability of the spaces between points. By anchoring the output of the algorithm so that Poisson point processes score on average 0, and perfect lattices score 10, we constructed an absolute interval scale of order. We demonstrated its utility in biology by quantifying the order of the Drosophila dorsal thorax epithelium during development. The psychophysical scaling method used relies on the comparison of stimuli with similar levels of order yielding a discrimination-based scale. As with other perceptual dimensions, an interesting question is whether supra-threshold perceptual differences are consistent with this scale. To test that we collected discrimination data, and data based on comparison of perceptual differences. Although the judgements of perceptual differences were found to be consistent with an interval scale, like the discrimination judgements, no common interval scale that could predict both sets of data was possible. Point patterns are commonly displayed as arrangements of dots. To examine how presentation parameters (dot size, dot numbers, and pattern area) affect discrimination, we collected discrimination data for ten presentation conditions. We found that discrimination performance depends on the ratio ‘dot diameter / average dot spacing’

    Multigranularity Representations for Human Inter-Actions: Pose, Motion and Intention

    Get PDF
    Tracking people and their body pose in videos is a central problem in computer vision. Standard tracking representations reason about temporal coherence of detected people and body parts. They have difficulty tracking targets under partial occlusions or rare body poses, where detectors often fail, since the number of training examples is often too small to deal with the exponential variability of such configurations. We propose tracking representations that track and segment people and their body pose in videos by exploiting information at multiple detection and segmentation granularities when available, whole body, parts or point trajectories. Detections and motion estimates provide contradictory information in case of false alarm detections or leaking motion affinities. We consolidate contradictory information via graph steering, an algorithm for simultaneous detection and co-clustering in a two-granularity graph of motion trajectories and detections, that corrects motion leakage between correctly detected objects, while being robust to false alarms or spatially inaccurate detections. We first present a motion segmentation framework that exploits long range motion of point trajectories and large spatial support of image regions. We show resulting video segments adapt to targets under partial occlusions and deformations. Second, we augment motion-based representations with object detection for dealing with motion leakage. We demonstrate how to combine dense optical flow trajectory affinities with repulsions from confident detections to reach a global consensus of detection and tracking in crowded scenes. Third, we study human motion and pose estimation. We segment hard to detect, fast moving body limbs from their surrounding clutter and match them against pose exemplars to detect body pose under fast motion. We employ on-the-fly human body kinematics to improve tracking of body joints under wide deformations. We use motion segmentability of body parts for re-ranking a set of body joint candidate trajectories and jointly infer multi-frame body pose and video segmentation. We show empirically that such multi-granularity tracking representation is worthwhile, obtaining significantly more accurate multi-object tracking and detailed body pose estimation in popular datasets

    Three dimensional pattern recognition using feature-based indexing and rule-based search

    Full text link
    In flexible automated manufacturing, robots can perform routine operations as well as recover from atypical events, provided that process-relevant information is available to the robot controller. Real time vision is among the most versatile sensing tools, yet the reliability of machine-based scene interpretation can be questionable. The effort described here is focused on the development of machine-based vision methods to support autonomous nuclear fuel manufacturing operations in hot cells; This thesis presents a method to efficiently recognize 3D objects from 2D images based on feature-based indexing. Object recognition is the identification of correspondences between parts of a current scene and stored views of known objects, using chains of segments or indexing vectors. To create indexed object models, characteristic model image features are extracted during preprocessing. Feature vectors representing model object contours are acquired from several points of view around each object and stored. Recognition is the process of matching stored views with features or patterns detected in a test scene; Two sets of algorithms were developed, one for preprocessing and indexed database creation, and one for pattern searching and matching during recognition. At recognition time, those indexing vectors with the highest match probability are retrieved from the model image database, using a nearest neighbor search algorithm. The nearest neighbor search predicts the best possible match candidates. Extended searches are guided by a search strategy that employs knowledge-base (KB) selection criteria. The knowledge-based system simplifies the recognition process and minimizes the number of iterations and memory usage; Novel contributions include the use of a feature-based indexing data structure together with a knowledge base. Both components improve the efficiency of the recognition process by improved structuring of the database of object features and reducing data base size. This data base organization according to object features facilitates machine learning in the context of a knowledge-base driven recognition algorithm. Lastly, feature-based indexing permits the recognition of 3D objects based on a comparatively small number of stored views, further limiting the size of the feature database; Experiments with real images as well as synthetic images including occluded (partially visible) objects are presented. The experiments show almost perfect recognition with feature-based indexing, if the detected features in the test scene are viewed from the same angle as the view on which the model is based. The experiments also show that the knowledge base is a highly effective and efficient search tool recognition performance is improved without increasing the database size requirements. The experimental results indicate that feature-based indexing in combination with a knowledge-based system will be a useful methodology for automatic target recognition (ATR)
    • …
    corecore