66,910 research outputs found

    Saliency Driven Vasculature Segmentation with Infinite Perimeter Active Contour Model

    Get PDF
    Automated detection of retinal blood vessels plays an important role in advancing the understanding of the mechanism, diagnosis and treatment of cardiovascular disease and many systemic diseases, such as diabetic retinopathy and age-related macular degeneration. Here, we propose a new framework for precisely segmenting retinal vasculatures. The proposed framework consists of three steps. A non-local total variation model is adapted to the Retinex theory, which aims to address challenges presented by intensity inhomogeneities, and the relatively low contrast of thin vessels compared to the background. The image is then divided into superpixels, and a compactness-based saliency detection method is proposed to locate the object of interest. For better general segmentation performance, we then make use of a new infinite active contour model to segment the vessels in each superpixel. The proposed framework has wide applications, and the results show that our model outperforms its competitors

    Reconstructive Sparse Code Transfer for Contour Detection and Semantic Labeling

    Get PDF
    We frame the task of predicting a semantic labeling as a sparse reconstruction procedure that applies a target-specific learned transfer function to a generic deep sparse code representation of an image. This strategy partitions training into two distinct stages. First, in an unsupervised manner, we learn a set of generic dictionaries optimized for sparse coding of image patches. We train a multilayer representation via recursive sparse dictionary learning on pooled codes output by earlier layers. Second, we encode all training images with the generic dictionaries and learn a transfer function that optimizes reconstruction of patches extracted from annotated ground-truth given the sparse codes of their corresponding image patches. At test time, we encode a novel image using the generic dictionaries and then reconstruct using the transfer function. The output reconstruction is a semantic labeling of the test image. Applying this strategy to the task of contour detection, we demonstrate performance competitive with state-of-the-art systems. Unlike almost all prior work, our approach obviates the need for any form of hand-designed features or filters. To illustrate general applicability, we also show initial results on semantic part labeling of human faces. The effectiveness of our approach opens new avenues for research on deep sparse representations. Our classifiers utilize this representation in a novel manner. Rather than acting on nodes in the deepest layer, they attach to nodes along a slice through multiple layers of the network in order to make predictions about local patches. Our flexible combination of a generatively learned sparse representation with discriminatively trained transfer classifiers extends the notion of sparse reconstruction to encompass arbitrary semantic labeling tasks.Comment: to appear in Asian Conference on Computer Vision (ACCV), 201

    A Framework for Symmetric Part Detection in Cluttered Scenes

    Full text link
    The role of symmetry in computer vision has waxed and waned in importance during the evolution of the field from its earliest days. At first figuring prominently in support of bottom-up indexing, it fell out of favor as shape gave way to appearance and recognition gave way to detection. With a strong prior in the form of a target object, the role of the weaker priors offered by perceptual grouping was greatly diminished. However, as the field returns to the problem of recognition from a large database, the bottom-up recovery of the parts that make up the objects in a cluttered scene is critical for their recognition. The medial axis community has long exploited the ubiquitous regularity of symmetry as a basis for the decomposition of a closed contour into medial parts. However, today's recognition systems are faced with cluttered scenes, and the assumption that a closed contour exists, i.e. that figure-ground segmentation has been solved, renders much of the medial axis community's work inapplicable. In this article, we review a computational framework, previously reported in Lee et al. (2013), Levinshtein et al. (2009, 2013), that bridges the representation power of the medial axis and the need to recover and group an object's parts in a cluttered scene. Our framework is rooted in the idea that a maximally inscribed disc, the building block of a medial axis, can be modeled as a compact superpixel in the image. We evaluate the method on images of cluttered scenes.Comment: 10 pages, 8 figure

    Image and Volume Segmentation by Water Flow

    No full text
    A general framework for image segmentation is presented in this paper, based on the paradigm of water flow. The major water flow attributes like water pressure, surface tension and capillary force are defined in the context of force field generation and make the model adaptable to topological and geometrical changes. A flow-stopping image functional combining edge- and region-based forces is introduced to produce capability for both range and accuracy. The method is assessed qualitatively and quantitatively on synthetic and natural images. It is shown that the new approach can segment objects with complex shapes or weak-contrasted boundaries, and has good immunity to noise. The operator is also extended to 3-D, and is successfully applied to medical volume segmentation

    Grounding semantics in robots for Visual Question Answering

    Get PDF
    In this thesis I describe an operational implementation of an object detection and description system that incorporates in an end-to-end Visual Question Answering system and evaluated it on two visual question answering datasets for compositional language and elementary visual reasoning

    Traditional and new principles of perceptual grouping

    Get PDF
    Perceptual grouping refers to the process of determining which regions and parts of the visual scene belong together as parts of higher order perceptual units such as objects or patterns. In the early 20th century, Gestalt psychologists identified a set of classic grouping principles which specified how some image features lead to grouping between elements given that all other factors were held constant. Modern vision scientists have expanded this list to cover a wide range of image features but have also expanded the importance of learning and other non-image factors. Unlike early Gestalt accounts which were based largely on visual demonstrations, modern theories are often explicitly quantitative and involve detailed models of how various image features modulate grouping. Work has also been done to understand the rules by which different grouping principles integrate to form a final percept. This chapter gives an overview of the classic principles, modern developments in understanding them, and new principles and the evidence for them. There is also discussion of some of the larger theoretical issues about grouping such as at what stage of visual processing it occurs and what types of neural mechanisms may implement grouping principles

    Medical Image Segmentation by Water Flow

    No full text
    We present a new image segmentation technique based on the paradigm of water flow and apply it to medical images. The force field analogy is used to implement the major water flow attributes like water pressure, surface tension and adhesion so that the model achieves topological adaptability and geometrical flexibility. A new snake-like force functional combining edge- and region-based forces is introduced to produce capability for both range and accuracy. The method has been assessed qualitatively and quantitatively, and shows decent detection performance as well as ability to handle noise
    corecore