1,696 research outputs found

    Comparing Feature Detectors: A bias in the repeatability criteria, and how to correct it

    Full text link
    Most computer vision application rely on algorithms finding local correspondences between different images. These algorithms detect and compare stable local invariant descriptors centered at scale-invariant keypoints. Because of the importance of the problem, new keypoint detectors and descriptors are constantly being proposed, each one claiming to perform better (or to be complementary) to the preceding ones. This raises the question of a fair comparison between very diverse methods. This evaluation has been mainly based on a repeatability criterion of the keypoints under a series of image perturbations (blur, illumination, noise, rotations, homotheties, homographies, etc). In this paper, we argue that the classic repeatability criterion is biased towards algorithms producing redundant overlapped detections. To compensate this bias, we propose a variant of the repeatability rate taking into account the descriptors overlap. We apply this variant to revisit the popular benchmark by Mikolajczyk et al., on classic and new feature detectors. Experimental evidence shows that the hierarchy of these feature detectors is severely disrupted by the amended comparator.Comment: Fixed typo in affiliation

    Robust Object-Based Watermarking Using SURF Feature Matching and DFT Domain

    Get PDF
    In this paper we propose a robust object-based watermarking method, in which the watermark is embedded into the middle frequencies band of the Discrete Fourier Transform (DFT) magnitude of the selected object region, altogether with the Speeded Up Robust Feature (SURF) algorithm to allow the correct watermark detection, even if the watermarked image has been distorted. To recognize the selected object region after geometric distortions, during the embedding process the SURF features are estimated and stored in advance to be used during the detection process. In the detection stage, the SURF features of the distorted image are estimated and match them with the stored ones. From the matching result, SURF features are used to compute the Affine-transformation parameters and the object region is recovered. The quality of the watermarked image is measured using the Peak Signal to Noise Ratio (PSNR), Structural Similarity Index (SSIM) and the Visual Information Fidelity (VIF). The experimental results show the proposed method provides robustness against several geometric distortions, signal processing operations and combined distortions. The receiver operating characteristics (ROC) curves also show the desirable detection performance of the proposed method. The comparison with a previously reported methods based on different techniques is also provided

    A Review of Codebook Models in Patch-Based Visual Object Recognition

    No full text
    The codebook model-based approach, while ignoring any structural aspect in vision, nonetheless provides state-of-the-art performances on current datasets. The key role of a visual codebook is to provide a way to map the low-level features into a fixed-length vector in histogram space to which standard classifiers can be directly applied. The discriminative power of such a visual codebook determines the quality of the codebook model, whereas the size of the codebook controls the complexity of the model. Thus, the construction of a codebook is an important step which is usually done by cluster analysis. However, clustering is a process that retains regions of high density in a distribution and it follows that the resulting codebook need not have discriminant properties. This is also recognised as a computational bottleneck of such systems. In our recent work, we proposed a resource-allocating codebook, to constructing a discriminant codebook in a one-pass design procedure that slightly outperforms more traditional approaches at drastically reduced computing times. In this review we survey several approaches that have been proposed over the last decade with their use of feature detectors, descriptors, codebook construction schemes, choice of classifiers in recognising objects, and datasets that were used in evaluating the proposed methods

    Registration of Face Image Using Modified BRISK Feature Descriptor

    Get PDF
    Automatic face recognition is a hot area of research in the field of computer vision. Even though a lot of research have been done in this field, still researchers are unable to develop an algorithm which can detect the face images under all possible real time conditions. Automatic face recognition algorithms are used in a variety of applications such as surveillance, automatic tagging, and human-robot interaction etc. The main problem faced by researchers working with the above real time problems is the uncertainty about the pose of the detected face, i.e. if the pose of the sensed image differ from the images in the trained database most of the existing algorithms will fail. So researchers suggested and proved that the detection accuracy against pose variation can be improved if we considered image registration as a preprocessing step prior to face recognition. In this work, scale and rotation invariant features have been used for image registration. The important steps in feature based image registration are preprocessing, feature detection, feature matching, transformation estimation, and resampling. In this work, feature detectors and descriptors like SIFT, SURF, FAST, DAISY and BRISK are used. Among all these descriptors the BRISK descriptor performs the best. To avoid mismatches, using some threshold values, a modified BRISK descriptor has been proposed in this work. Modified BRISK descriptor performs best in terms of maximum matching as compared to other state of arts descriptors. The next step is to calculate the transformation model which is capable of transforming the coordinates of sensed image to coordinates of reference image. Some radial basis functions are used in this step to design the proper transformation function. In resampling step, we used bilinear interpolation to compute some pixels in the output image. A new algorithm is proposed in this work to find out the possible image pairs from the train database corresponds to the input image, for doing image registration. In this work, image registration algorithms are simulated in MATLAB with different detector-descriptor combination and affine transformation matrix. For measuring the similarity between registered output image and the reference image, SSIM index and mutual information is used

    WxBS: Wide Baseline Stereo Generalizations

    Full text link
    We have presented a new problem -- the wide multiple baseline stereo (WxBS) -- which considers matching of images that simultaneously differ in more than one image acquisition factor such as viewpoint, illumination, sensor type or where object appearance changes significantly, e.g. over time. A new dataset with the ground truth for evaluation of matching algorithms has been introduced and will be made public. We have extensively tested a large set of popular and recent detectors and descriptors and show than the combination of RootSIFT and HalfRootSIFT as descriptors with MSER and Hessian-Affine detectors works best for many different nuisance factors. We show that simple adaptive thresholding improves Hessian-Affine, DoG, MSER (and possibly other) detectors and allows to use them on infrared and low contrast images. A novel matching algorithm for addressing the WxBS problem has been introduced. We have shown experimentally that the WxBS-M matcher dominantes the state-of-the-art methods both on both the new and existing datasets.Comment: Descriptor and detector evaluation expande
    corecore