46 research outputs found

    Analyzing scenery images by monotonic tree

    Full text link
    Content-based image retrieval (CBIR) has been an active research area in the last ten years, and a variety of techniques have been developed. However, retrieving images on the basis of low-level features has proven unsatisfactory, and new techniques are needed to support high-level queries. Research efforts are needed to bridge the gap between high-level semantics and low-level features. In this paper, we present a novel approach to support semantics-based image retrieval. Our approach is based on the monotonic tree, a derivation of the contour tree for use with discrete data. The structural elements of an image are modeled as branches (or subtrees) of the monotonic tree. These structural elements are classified and clustered on the basis of such properties as color, spatial location, harshness and shape. Each cluster corresponds to some semantic feature. This scheme is applied to the analysis and retrieval of scenery images. Comparisons of experimental results of this approach with conventional techniques using low-level features demonstrate the effectiveness of our approach.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/42315/1/30080495.pd

    Automatic Annotation and Retrieval of Images

    Full text link
    Although a variety of techniques have been developed for content-based image retrieval (CBIR), automatic image retrieval by semantics still remains a challenging problem. We propose a novel approach for semantics-based image annotation and retrieval. Our approach is based on the monotonic tree model. The branches of the monotonic tree of an image, termed as structural elements , are classified and clustered based on their low level features such as color, spatial location, coarseness, and shape. Each cluster corresponds to some semantic feature. The category keywords indicating the semantic features are automatically annotated to the images. Based on the semantic features extracted from images, high-level (semantics-based) querying and browsing of images can be achieved. We apply our scheme to analyze scenery features. Experiments show that semantic features, such as sky, building, trees, water wave, placid water, and ground, can be effectively retrieved and located in images.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/41397/1/11280_2004_Article_5122908.pd

    A Model for Weighting Image Objects in Home Photographs

    No full text
    International audienceThe paper presents a contribution to image indexing consisting in a weighting model for visible objects - or image objects - in home photographs. To improve its effectiveness this weighting model has been designed according to human perception criteria about what is estimated as important in photographs. Four basic hypotheses related to human perception are presented, and their validity is estimated as compared to actual observations from a user study. Finally a formal definition of this weighting model is presented and its consistence with the user study is evaluated

    On effects of visual query complexity

    Get PDF

    Multimedia indexing and fast retrieval based on a vote system

    Full text link

    On the Performance Improvement of Iris Biometric System

    Get PDF
    Iris is an established biometric modality with many practical applications. Its performance is influenced by noise, database size, and feature representation. This thesis focusses on mitigating these challenges by efficiently characterising iris texture,developing multi-unit iris recognition, reducing the search space of large iris databases, and investigating if iris pattern change over time.To suitably characterise texture features of iris, Scale Invariant Feature Transform (SIFT) is combined with Fourier transform to develop a keypoint descriptor-F-SIFT. Proposed F-SIFT is invariant to transformation, illumination, and occlusion along with strong texture description property. For pairing the keypoints from gallery and probe iris images, Phase-Only Correlation (POC) function is used. The use of phase information reduces the wrong matches generated using SIFT. Results demonstrate the effectiveness of F-SIFT over existing keypoint descriptors.To perform the multi-unit iris fusion, a novel classifier is proposed known as Incremental Granular Relevance Vector Machine (iGRVM) that incorporates incremental and granular learning into RVM. The proposed classifier by design is scalable and unbiased which is particularly suitable for biometrics. The match scores from individual units of iris are passed as an input to the corresponding iGRVM classifier, and the posterior probabilities are combined using weighted sum rule. Experimentally, it is shown that the performance of multi-unit iris recognition improves over single unit iris. For search space reduction, local feature based indexing approaches are developed using multi-dimensional trees. Such features extracted from annular iris images are used to index the database using k-d tree. To handle the scalability issue of k-d tree, k-d-b tree based indexing approach is proposed. Another indexing approach using R-tree is developed to minimise the indexing errors. For retrieval, hybrid coarse-to-fine search strategy is proposed. It is inferred from the results that unification of hybrid search with R-tree significantly improves the identification performance. Iris is assumed to be stable over time. Recently, researchers have reported that false rejections increase over the period of time which in turn degrades the performance. An empirical investigation has been made on standard iris aging databases to find whether iris patterns change over time. From the results, it is found that the rejections are primarily due to the presence of other covariates such as blur, noise, occlusion, pupil dilation, and not due to agin

    Information fusion in content based image retrieval: A comprehensive overview

    Get PDF
    An ever increasing part of communication between persons involve the use of pictures, due to the cheap availability of powerful cameras on smartphones, and the cheap availability of storage space. The rising popularity of social networking applications such as Facebook, Twitter, Instagram, and of instant messaging applications, such as WhatsApp, WeChat, is the clear evidence of this phenomenon, due to the opportunity of sharing in real-time a pictorial representation of the context each individual is living in. The media rapidly exploited this phenomenon, using the same channel, either to publish their reports, or to gather additional information on an event through the community of users. While the real-time use of images is managed through metadata associated with the image (i.e., the timestamp, the geolocation, tags, etc.), their retrieval from an archive might be far from trivial, as an image bears a rich semantic content that goes beyond the description provided by its metadata. It turns out that after more than 20 years of research on Content-Based Image Retrieval (CBIR), the giant increase in the number and variety of images available in digital format is challenging the research community. It is quite easy to see that any approach aiming at facing such challenges must rely on different image representations that need to be conveniently fused in order to adapt to the subjectivity of image semantics. This paper offers a journey through the main information fusion ingredients that a recipe for the design of a CBIR system should include to meet the demanding needs of users
    corecore