1,380 research outputs found

    TRACKING THE URBAN CHAMELEON – TOWARDS A HYBRID CHANGE DETECTION OF GRAFFITI

    Get PDF
    Colourful and ever-changing: Graffiti can be considered the urban chameleon skin. At the Donaukanal (Eng. Danube Channel), Vienna's central waterway and one of the largest and most active graffiti-scapes worldwide, this metaphor applies like hardly anywhere else. Every day a multitude of graffiti is destroyed by the creation of new works. Recently, efforts have been made to mitigate this constant loss of cultural heritage along the Donaukanal by systematically documenting the graffiti, mainly using photography and photogrammetry. However, keeping track of the newly added works is very time-consuming and often like finding needles in a haystack, considering the large extent and high volatility of the monitored area. Thus, an automated graffiti change detection would significantly reduce the effort and avoid overlooking graffiti.This contribution outlines the main challenges in image-based change detection for cultural heritage and proposes a hybrid graffiti change detection method. The investigated method exploits and combines an established pixel-based change detection algorithm, the Iteratively Multivariate Alteration Detection, with a novel descriptor-based method. The latter relies on image features, rather than pixels as analysis unit and can robustly filter false alarms from the high-performing but noise-prone pixel-based approach. Overall, the results indicate that the proposed method can largely automate image-based change detection of graffiti-scapes. It can uncover graffiti-related changes and robustly distinguish them from other image differences such as shadows but tends to overlook small-scale graffiti, indicating the need for further fine-tuning.</p

    Superpixel-based Two-view Deterministic Fitting for Multiple-structure Data

    Full text link
    This paper proposes a two-view deterministic geometric model fitting method, termed Superpixel-based Deterministic Fitting (SDF), for multiple-structure data. SDF starts from superpixel segmentation, which effectively captures prior information of feature appearances. The feature appearances are beneficial to reduce the computational complexity for deterministic fitting methods. SDF also includes two original elements, i.e., a deterministic sampling algorithm and a novel model selection algorithm. The two algorithms are tightly coupled to boost the performance of SDF in both speed and accuracy. Specifically, the proposed sampling algorithm leverages the grouping cues of superpixels to generate reliable and consistent hypotheses. The proposed model selection algorithm further makes use of desirable properties of the generated hypotheses, to improve the conventional fit-and-remove framework for more efficient and effective performance. The key characteristic of SDF is that it can efficiently and deterministically estimate the parameters of model instances in multi-structure data. Experimental results demonstrate that the proposed SDF shows superiority over several state-of-the-art fitting methods for real images with single-structure and multiple-structure data.Comment: Accepted by European Conference on Computer Vision (ECCV

    Scale Invariant Interest Points with Shearlets

    Full text link
    Shearlets are a relatively new directional multi-scale framework for signal analysis, which have been shown effective to enhance signal discontinuities such as edges and corners at multiple scales. In this work we address the problem of detecting and describing blob-like features in the shearlets framework. We derive a measure which is very effective for blob detection and closely related to the Laplacian of Gaussian. We demonstrate the measure satisfies the perfect scale invariance property in the continuous case. In the discrete setting, we derive algorithms for blob detection and keypoint description. Finally, we provide qualitative justifications of our findings as well as a quantitative evaluation on benchmark data. We also report an experimental evidence that our method is very suitable to deal with compressed and noisy images, thanks to the sparsity property of shearlets

    Large-scale image collection cleansing, summarization and exploration

    Get PDF
    A perennially interesting topic in the research field of large scale image collection organization is how to effectively and efficiently conduct the tasks of image cleansing, summarization and exploration. The primary objective of such an image organization system is to enhance user exploration experience with redundancy removal and summarization operations on large-scale image collection. An ideal system is to discover and utilize the visual correlation among the images, to reduce the redundancy in large-scale image collection, to organize and visualize the structure of large-scale image collection, and to facilitate exploration and knowledge discovery. In this dissertation, a novel system is developed for exploiting and navigating large-scale image collection. Our system consists of the following key components: (a) junk image filtering by incorporating bilingual search results; (b) near duplicate image detection by using a coarse-to-fine framework; (c) concept network generation and visualization; (d) image collection summarization via dictionary learning for sparse representation; and (e) a multimedia practice of graffiti image retrieval and exploration. For junk image filtering, bilingual image search results, which are adopted for the same keyword-based query, are integrated to automatically identify the clusters for the junk images and the clusters for the relevant images. Within relevant image clusters, the results are further refined by removing the duplications under a coarse-to-fine structure. The duplicate pairs are detected with both global feature (partition based color histogram) and local feature (CPAM and SIFT Bag-of-Word model). The duplications are detected and removed from the data collection to facilitate further exploration and visual correlation analysis. After junk image filtering and duplication removal, the visual concepts are further organized and visualized by the proposed concept network. An automatic algorithm is developed to generate such visual concept network which characterizes the visual correlation between image concept pairs. Multiple kernels are combined and a kernel canonical correlation analysis algorithm is used to characterize the diverse visual similarity contexts between the image concepts. The FishEye visualization technique is implemented to facilitate the navigation of image concepts through our image concept network. To better assist the exploration of large scale data collection, we design an efficient summarization algorithm to extract representative examplars. For this collection summarization task, a sparse dictionary (a small set of the most representative images) is learned to represent all the images in the given set, e.g., such sparse dictionary is treated as the summary for the given image set. The simulated annealing algorithm is adopted to learn such sparse dictionary (image summary) by minimizing an explicit optimization function. In order to handle large scale image collection, we have evaluated both the accuracy performance of the proposed algorithms and their computation efficiency. For each of the above tasks, we have conducted experiments on multiple public available image collections, such as ImageNet, NUS-WIDE, LabelMe, etc. We have observed very promising results compared to existing frameworks. The computation performance is also satisfiable for large-scale image collection applications. The original intention to design such a large-scale image collection exploration and organization system is to better service the tasks of information retrieval and knowledge discovery. For this purpose, we utilize the proposed system to a graffiti retrieval and exploration application and receive positive feedback

    Rotation Invariant on Harris Interest Points for Exposing Image Region Duplication Forgery

    Get PDF
    Nowadays, image forgery has become common because only an editing package software and a digital camera are required to counterfeit an image. Various fraud detection systems have been developed in accordance with the requirements of numerous applications and to address different types of image forgery. However, image fraud detection is a complicated process given that is necessary to identify the image processing tools used to counterfeit an image. Here, we describe recent developments in image fraud detection. Conventional techniques for detecting duplication forgeries have difficulty in detecting postprocessing falsification, such as grading and joint photographic expert group compression. This study proposes an algorithm that detects image falsification on the basis of Hessian features

    FPGA-based module for SURF extraction

    Get PDF
    We present a complete hardware and software solution of an FPGA-based computer vision embedded module capable of carrying out SURF image features extraction algorithm. Aside from image analysis, the module embeds a Linux distribution that allows to run programs specifically tailored for particular applications. The module is based on a Virtex-5 FXT FPGA which features powerful configurable logic and an embedded PowerPC processor. We describe the module hardware as well as the custom FPGA image processing cores that implement the algorithm's most computationally expensive process, the interest point detection. The module's overall performance is evaluated and compared to CPU and GPU based solutions. Results show that the embedded module achieves comparable disctinctiveness to the SURF software implementation running in a standard CPU while being faster and consuming significantly less power and space. Thus, it allows to use the SURF algorithm in applications with power and spatial constraints, such as autonomous navigation of small mobile robots
    • …
    corecore