48,301 research outputs found

    Range Image Segmentation for 3-D Object Recognition

    Get PDF
    Three dimensional scene analysis in an unconstrained and uncontrolled environment is the ultimate goal of computer vision. Explicit depth information about the scene is of tremendous help in segmentation and recognition of objects. Range image interpretation with a view of obtaining low-level features to guide mid-level and high-level segmentation and recognition processes is described. No assumptions about the scene are made and algorithms are applicable to any general single viewpoint range image. Low-level features like step edges and surface characteristics are extracted from the images and segmentation is performed based on individual features as well as combination of features. A high level recognition process based on superquadric fitting is described to demonstrate the usefulness of initial segmentation based on edges. A classification algorithm based on surface curvatures is used to obtain initial segmentation of the scene. Objects segmented using edge information are then classified using surface curvatures. Various applications of surface curvatures in mid and high level recognition processes are discussed. These include surface reconstruction, segmentation into convex patches and detection of smooth edges. Algorithms are run on real range images and results are discussed in detail

    A Technique for Optimal Selection of Segmentation Scale Parameters for Object-oriented Classification of Urban Scenes

    Get PDF
    Multi-scale image segmentation produces high level object features at more than one level, compared to single scale segmentation. Objects generated from this type of segmentation hold additional attributes such as mean values per spectral band, distances to neighbouring objects, size, and texture, as well as shape characteristics. However, the accuracy of these high level features depends on the choice of segmentation scale parameters. Several studies have investigated techniques for scale parameter selection. These proposed approaches do not consider the different objects’ size variability found in complex scenes such as urban scene as they rely upon arbitrary object size measures, introducing instability errors when computing image variances. A technique to select optimal segmentation scale parameters based on image variance and spatial autocorrelation is presented in this paper. Optimal scales satisfy simultaneously the conditions of low object internal variance and high inter-segments spatial autocorrelation. Applied on three Cape Town urban scenes, the technique produced visually promising results that would  improve object extraction over urban areas.Key words: segmentation, object oriented classification, object’s variance, spatial autocorrelation, objective function, Moran’s index

    Multi-Resolution Feature Embedded Level Set Model for Crosshatched Texture Segmentation

    Get PDF
    In image processing applications, texture is the most important element utilized by human visual systems for distinguishing dissimilar objects in a scene. In this research article, a variational model based on the level set is implemented for crosshatched texture segmentation. In this study, the proposed model’s performance is validated on the Brodatz texture dataset. The cross-hatched texture segmentation in the lower resolution texture images is difficult, due to the computational and memory requirements. The aforementioned issue has been resolved by implementing a variational model based on the level set that enables efficient segmentation in both low and high-resolution images with automatic selection of the filter size. In the proposed model, the multi-resolution feature obtained from the frequency domain filters enhances the dissimilarity between the regions of crosshatched textures that have low-intensity variations. Then, the resultant images are integrated with a level set-based active contour model that addresses the segmentation of crosshatched texture images. The noise added during the segmentation process is eliminated by morphological processing. The experiments conducted on the Brodatz texture dataset demonstrated the effectiveness of the proposed model, and the obtained results are validated in terms of Intersection over the Union (IoU) index, accuracy, precision, f1-score and recall. The extensive experimental investigation shows that the proposed model effectively segments the region of interest in close correspondence with the original image. The proposed segmentation model with a multi-support vector machine has achieved a classification accuracy of 99.82%, which is superior to the comparative model (modified convolutional neural network with whale optimization algorithm). The proposed model almost showed a 0.11% improvement in classification accuracy related to the existing mode

    Bottom-up Object Segmentation for Visual Recognition

    Get PDF
    Automatic recognition and segmentation of objects in images is a central open problem in computer vision. Most previous approaches have pursued either sliding-window object detection or dense classification of overlapping local image patches. Differently, the framework introduced in this thesis attempts to identify the spatial extent of objects prior to recognition, using bottom-up computational processes and mid-level selection cues. After a set of plausible object hypotheses is identified, a sequential recognition process is executed, based on continuous estimates of the spatial overlap between the image segment hypotheses and each putative class. The object hypotheses are represented as figure-ground segmentations, and are extracted automatically, without prior knowledge of the properties of individual object classes, by solving a sequence of constrained parametric min-cut problems (CPMC) on a regular image grid. It is show that CPMC significantly outperforms the state of the art for low-level segmentation in the PASCAL VOC 2009 and 2010 datasets. Results beyond the current state of the art for image classification, object detection and semantic segmentation are also demonstrated in a number of challenging datasets including Caltech-101, ETHZ-Shape as well as PASCAL VOC 2009-11. These results suggest that a greater emphasis on grouping and image organization may be valuable for making progress in high-level tasks such as object recognition and scene understanding

    Driving Scene Perception Network: Real-time Joint Detection, Depth Estimation and Semantic Segmentation

    Full text link
    As the demand for enabling high-level autonomous driving has increased in recent years and visual perception is one of the critical features to enable fully autonomous driving, in this paper, we introduce an efficient approach for simultaneous object detection, depth estimation and pixel-level semantic segmentation using a shared convolutional architecture. The proposed network model, which we named Driving Scene Perception Network (DSPNet), uses multi-level feature maps and multi-task learning to improve the accuracy and efficiency of object detection, depth estimation and image segmentation tasks from a single input image. Hence, the resulting network model uses less than 850 MiB of GPU memory and achieves 14.0 fps on NVIDIA GeForce GTX 1080 with a 1024x512 input image, and both precision and efficiency have been improved over combination of single tasks.Comment: 9 pages, 7 figures, WACV'1
    • …
    corecore