78 research outputs found
Object representation and recognition
One of the primary functions of the human visual system is object recognition, an ability that allows us to relate the visual stimuli falling on our retinas to our knowledge of the world. For example, object recognition allows you to use knowledge of what an apple looks like to find it in the supermarket, to use knowledge of what a shark looks like to swim in th
Multi-Scale Free-Form Surface Description and Curvature Estimation
A novel technique for multi-scale smoothing of a free-form 3-D surface is presented. Complete triangulated models of 3-D objects are constructed at our center [4] and using a local parametrization technique, are then smoothed using a 2-D Gaussian filter. Our method for local parametrization makes use of semigeodesic coordinates as a natural and efficient way of sampling the local surface shape. The smoothing eliminates the surface noise together with high curvature regions such as sharp edges, therefore, sharp corners become rounded as the object is smoothed iteratively. Our technique for free-form 3-D multi-scale surface smoothing is independent of the underlying triangulation. It is also argued that the proposed technique is preferrable to volumetric smoothing or level set methods since it is applicable to incomplete surface data which occurs during occlusion. The technique was applied to simple and complex 3-D objects and the results are presented here
Part-based Grouping and Recognition: A Model-Guided Approach
Institute of Perception, Action and BehaviourThe recovery of generic solid parts is a fundamental step towards the realization of
general-purpose vision systems. This thesis investigates issues in grouping, segmentation and recognition of parts from two-dimensional edge images.
A new paradigm of part-based grouping of features is introduced that bridges the classical grouping and model-based approaches with the purpose of directly recovering
parts from real images, and part-like models are used that both yield low theoretical
complexity and reliably recover part-plausible groups of features. The part-like models
used are statistical point distribution models, whose training set is built using random
deformable superellipse.
The computational approach that is proposed to perform model-guided part-based
grouping consists of four distinct stages.
In the first stage, codons, contour portions of similar curvature, are extracted from the
raw edge image. They are considered to be indivisible image features because they
have the desirable property of belonging either to single parts or joints.
In the second stage, small seed groups (currently pairs, but further extension are proposed) of codons are found that give enough structural information for part hypotheses
to be created. The third stage consists in initialising and pre-shaping the models to
all the seed groups and then performing a full fitting to a large neighbourhood of the
pre-shaped model. The concept of pre-shaping to a few significant features is a relatively new concept in deformable model fitting that has helped to dramatically increase
robustness. The initialisations of the part models to the seed groups is performed by
the first direct least-square ellipse fitting algorithm, which has been jointly discovered
during this research; a full theoretical proof of the method is provided.
The last stage pertains to the global filtering of all the hypotheses generated by the previous stages according to the Minimum Description Length criterion: the small number
of grouping hypotheses that survive this filtering stage are the most economical representation of the image in terms of the part-like models. The filtering is performed by
the maximisation of a boolean quadratic function by a genetic algorithm, which has
resulted in the best trade-off between speed and robustness.
Finally, images of parts can have a pronounced 3D structure, with ends or sides clearly
visible. In order to recover this important information, the part-based grouping method
is extended by employing parametrically deformable aspects models which, starting
from the initial position provided by the previous stages, are fitted to the raw image
by simulated annealing. These models are inspired by deformable superquadrics but
are built by geometric construction, which render them two order of magnitudes faster
to generate than in previous works.
A large number of experiments is provided that validate the approach and, since several
new issues have been opened by it, some future work is proposed
Superquadric representation of scenes from multi-view range data
Object representation denotes representing three-dimensional (3D) real-world objects with known graphic or mathematic primitives recognizable to computers. This research has numerous applications for object-related tasks in areas including computer vision, computer graphics, reverse engineering, etc. Superquadrics, as volumetric and parametric models, have been selected to be the representation primitives throughout this research. Superquadrics are able to represent a large family of solid shapes by a single equation with only a few parameters. This dissertation addresses superquadric representation of multi-part objects and multiobject scenes. Two issues motivate this research. First, superquadric representation of multipart objects or multi-object scenes has been an unsolved problem due to the complex geometry of objects. Second, superquadrics recovered from single-view range data tend to have low confidence and accuracy due to partially scanned object surfaces caused by inherent occlusions. To address these two problems, this dissertation proposes a multi-view superquadric representation algorithm. By incorporating both part decomposition and multi-view range data, the proposed algorithm is able to not only represent multi-part objects or multi-object scenes, but also achieve high confidence and accuracy of recovered superquadrics. The multi-view superquadric representation algorithm consists of (i) initial superquadric model recovery from single-view range data, (ii) pairwise view registration based on recovered superquadric models, (iii) view integration, (iv) part decomposition, and (v) final superquadric fitting for each decomposed part. Within the multi-view superquadric representation framework, this dissertation proposes a 3D part decomposition algorithm to automatically decompose multi-part objects or multiobject scenes into their constituent single parts consistent with human visual perception. Superquadrics can then be recovered for each decomposed single-part object. The proposed part decomposition algorithm is based on curvature analysis, and includes (i) Gaussian curvature estimation, (ii) boundary labeling, (iii) part growing and labeling, and (iv) post-processing. In addition, this dissertation proposes an extended view registration algorithm based on superquadrics. The proposed view registration algorithm is able to handle deformable superquadrics as well as 3D unstructured data sets. For superquadric fitting, two objective functions primarily used in the literature have been comprehensively investigated with respect to noise, viewpoints, sample resolutions, etc. The objective function proved to have better performance has been used throughout this dissertation. In summary, the three algorithms (contributions) proposed in this dissertation are generic and flexible in the sense of handling triangle meshes, which are standard surface primitives in computer vision and graphics. For each proposed algorithm, the dissertation presents both theory and experimental results. The results demonstrate the efficiency of the algorithms using both synthetic and real range data of a large variety of objects and scenes. In addition, the experimental results include comparisons with previous methods from the literature. Finally, the dissertation concludes with a summary of the contributions to the state of the art in superquadric representation, and presents possible future extensions to this research
Integration of Quantitative and Qualitative Techniques for Deformable Model Fitting from Orthographic, Perspective, and Stereo Projections
In this paper, we synthesize a new approach to 3-D object shape recovery by integrating qualitative shape recovery techniques and quantitative physics based shape estimation techniques. Specifically, we first use qualitative shape recovery and recognition techniques to provide strong fitting constraints on physics-based deformable model recovery techniques. Secondly, we extend our previously developed technique of fitting deformable models to occluding image contours to the case of image data captured under general orthographic, perspective, and stereo projections
Component-wise modeling of articulated objects
We introduce a novel framework for modeling articulated objects based on the aspects of their components. By decomposing the object into components, we divide the problem in smaller modeling tasks. After obtaining 3D models for each component aspect by employing a shape deformation paradigm, we merge them together, forming the object components. The final model is obtained by assembling the components using an optimization scheme which fits the respective 3D models to the corresponding apparent contours in a reference pose. The results suggest that our approach can produce realistic 3D models of articulated objects in reasonable time
Complex scene modeling and segmentation with deformable simplex meshes
In this thesis we present a system for 3D reconstruction and segmentation of complex real world scenes. The input to the system is an unstructured cloud of 3D points. The output is a 3D model for each object in the scene. The system starts with a model that encloses the input point cloud. A deformation process is applied to the initial model so it gets close to the point cloud in terms of distance, geometry and topology. Once the deformation stops the model is analyzed to check if more than one object is present in the point cloud. If necessary a segmentation process splits the model into several parts that correspond to each object in the scene. Using this segmented model the point cloud is also segmented. Each resulting sub-cloud is treated as a new input to the system. If, after the deformation process, the model is not segmented a refinement process improves the objective and subjective quality of the model by concentrating vertices around high curvature areas. The simplex mesh reconstruction algorithm was modified and extended to suit our application. A novel segmentation algorithm was designed to be applied on the simplex mesh. We test the system with synthetic and real data obtained from single objects, simple. and complex scenes. In the case of the synthetic data different levels of noise are added to examine the performance of the system. The results show that the systems performs well for either of the three cases and also in the presence of low levels of noise
Recognition by Functional Parts
(Also cross-referenced as CAR-TR-703)
We present an approach to function-based object recognition that
reasons about the functionality of an object's intuitive parts. We extend
the popular "recognition by parts" shape recognition framework to support
"recognition by functional parts", by com bining a set of functional
primitives and their relations with a set of abstract volumetric shape
primitives and their relations. Previous approaches have relied on more
global object features, often ignoring the problem of object segmentation
and thereby restricting themselves to range images of unoccluded scenes.
We show how these shape primitives and relations can be easily recovered
from superquadric ellipsoids which, in turn, can be recovered from
either range or intensity images of occluded scenes. Furthermore, the
proposed framework supports both unexpected (bottom-up) object
recognition and expected (top-down) object recognition. We demonstrate
the approach on a simple domain by recognizing a restricted class of
hand-tools from 2-D images
Vertebral body segmentation with GrowCut: Initial experience, workflow and practical application
In this contribution, we used the GrowCut segmentation algorithm publicly
available in three-dimensional Slicer for three-dimensional segmentation of
vertebral bodies. To the best of our knowledge, this is the first time that the
GrowCut method has been studied for the usage of vertebral body segmentation.
In brief, we found that the GrowCut segmentation times were consistently less
than the manual segmentation times. Hence, GrowCut provides an alternative to a
manual slice-by-slice segmentation process.Comment: 10 page
- …