940 research outputs found

    Visual saliency computation for image analysis

    Full text link
    Visual saliency computation is about detecting and understanding salient regions and elements in a visual scene. Algorithms for visual saliency computation can give clues to where people will look in images, what objects are visually prominent in a scene, etc. Such algorithms could be useful in a wide range of applications in computer vision and graphics. In this thesis, we study the following visual saliency computation problems. 1) Eye Fixation Prediction. Eye fixation prediction aims to predict where people look in a visual scene. For this problem, we propose a Boolean Map Saliency (BMS) model which leverages the global surroundedness cue using a Boolean map representation. We draw a theoretic connection between BMS and the Minimum Barrier Distance (MBD) transform to provide insight into our algorithm. Experiment results show that BMS compares favorably with state-of-the-art methods on seven benchmark datasets. 2) Salient Region Detection. Salient region detection entails computing a saliency map that highlights the regions of dominant objects in a scene. We propose a salient region detection method based on the Minimum Barrier Distance (MBD) transform. We present a fast approximate MBD transform algorithm with an error bound analysis. Powered by this fast MBD transform algorithm, our method can run at about 80 FPS and achieve state-of-the-art performance on four benchmark datasets. 3) Salient Object Detection. Salient object detection targets at localizing each salient object instance in an image. We propose a method using a Convolutional Neural Network (CNN) model for proposal generation and a novel subset optimization formulation for bounding box filtering. In experiments, our subset optimization formulation consistently outperforms heuristic bounding box filtering baselines, such as Non-maximum Suppression, and our method substantially outperforms previous methods on three challenging datasets. 4) Salient Object Subitizing. We propose a new visual saliency computation task, called Salient Object Subitizing, which is to predict the existence and the number of salient objects in an image using holistic cues. To this end, we present an image dataset of about 14K everyday images which are annotated using an online crowdsourcing marketplace. We show that an end-to-end trained CNN subitizing model can achieve promising performance without requiring any localization process. A method is proposed to further improve the training of the CNN subitizing model by leveraging synthetic images. 5) Top-down Saliency Detection. Unlike the aforementioned tasks, top-down saliency detection entails generating task-specific saliency maps. We propose a weakly supervised top-down saliency detection approach by modeling the top-down attention of a CNN image classifier. We propose Excitation Backprop and the concept of contrastive attention to generate highly discriminative top-down saliency maps. Our top-down saliency detection method achieves superior performance in weakly supervised localization tasks on challenging datasets. The usefulness of our method is further validated in the text-to-region association task, where our method provides state-of-the-art performance using only weakly labeled web images for training

    Advanced Visual Computing for Image Saliency Detection

    Get PDF
    Saliency detection is a category of computer vision algorithms that aims to filter out the most salient object in a given image. Existing saliency detection methods can generally be categorized as bottom-up methods and top-down methods, and the prevalent deep neural network (DNN) has begun to show its applications in saliency detection in recent years. However, the challenges in existing methods, such as problematic pre-assumption, inefficient feature integration and absence of high-level feature learning, prevent them from superior performances. In this thesis, to address the limitations above, we have proposed multiple novel models with favorable performances. Specifically, we first systematically reviewed the developments of saliency detection and its related works, and then proposed four new methods, with two based on low-level image features, and two based on DNNs. The regularized random walks ranking method (RR) and its reversion-correction-improved version (RCRR) are based on conventional low-level image features, which exhibit higher accuracy and robustness in extracting the image boundary based foreground / background queries; while the background search and foreground estimation (BSFE) and dense and sparse labeling (DSL) methods are based on DNNs, which have shown their dominant advantages in high-level image feature extraction, as well as the combined strength of multi-dimensional features. Each of the proposed methods is evaluated by extensive experiments, and all of them behave favorably against the state-of-the-art, especially the DSL method, which achieves remarkably higher performance against sixteen state-of-the-art methods (including ten conventional methods and six learning based methods) on six well-recognized public datasets. The successes of our proposed methods reveal more potential and meaningful applications of saliency detection in real-life computer vision tasks

    Melhorias na segmentação de pele humana em imagens digitais

    Get PDF
    Orientador: Hélio PedriniDissertação (mestrado) - Universidade Estadual de Campinas, Instituto de ComputaçãoResumo: Segmentação de pele humana possui diversas aplicações nas áreas de visão computacional e reconhecimento de padrões, cujo propósito principal é distinguir regiões de pele e não pele em imagens. Apesar do elevado número de métodos disponíveis na literatura, a segmentação de pele com precisão ainda é uma tarefa desafiadora. Muitos métodos contam somente com a informação de cor, o que não discrimina completamente as regiões da imagem devido a variações nas condições de iluminação e à ambiguidade entre a cor da pele e do plano de fundo. Dessa forma, há ainda a demanda em melhorar a segmentação. Este trabalho apresenta três contribuições com respeito a essa necessidade. A primeira é um método autocontido para segmentação adaptativa de pele que faz uso de análise espacial para produzir regiões nas quais a cor da pele é estimada e, dessa forma, ajusta o padrão da cor para uma imagem em particular. A segunda é a introdução da detecção de saliência para, combinada com detectores de pele baseados em cor, realizar a remoção do plano de fundo, o que elimina muitas regiões de não pele. A terceira é uma melhoria baseada em textura utilizando superpixels para capturar energia de regiões na imagem filtrada, que é então utilizada para caracterizar regiões de não pele e assim eliminar a ambiguidade da cor adicionando um segundo voto. Resultados experimentais obtidos em bases de dados públicas comprovam uma melhoria significativa nos métodos propostos para segmentação de pele humana em comparação com abordagens disponíveis na literaturaAbstract: Human skin segmentation has several applications on computer vision and pattern recognition fields, whose main purpose is to distinguish skin and non-skin regions. Despite the large number of methods available in the literature, accurate skin segmentation is still a challenging task. Many methods rely only on color information, which does not completely discriminate the image regions due to variations in lighting conditions and ambiguity between skin and background color. Therefore, there is still demand to improve the segmentation process. Three main contributions toward this need are presented in this work. The first is a self-contained method for adaptive skin segmentation that makes use of spatial analysis to produce regions from which the overall skin color can be estimated and such that the color model is adjusted to a particular image. The second is the combination of saliency detection with color skin segmentation, which performs a background removal to eliminate non-skin regions. The third is a texture-based improvement using superpixels to capture energy of regions in the filtered image, employed to characterize non-skin regions and thus eliminate color ambiguity adding a second vote. Experimental results on public data sets demonstrate a significant improvement of the proposed methods for human skin segmentation over state-of-the-art approachesMestradoCiência da ComputaçãoMestre em Ciência da Computaçã
    corecore