91,690 research outputs found

    Algorithm Selection for Edge Detection in Satellite Images by Neutrosophic WASPAS Method

    Get PDF
    Nowadays, integrated land management is generally governed by the principles of sustainability. Land use management usually is grounded in satellite image information. The detection and monitoring of areas of interest in satellite images is a difficult task. We propose a new methodology for the adaptive selection of edge detection algorithms using visual features of satellite images and the multi-criteria decision-making (MCDM) method. It is not trivial to select the most appropriate method for the chosen satellite images as there is no proper algorithm for all cases as it depends on many factors, like acquisition and content of the raster images, visual features of real-world images, and humans’ visual perception. The edge detection algorithms were ranked according to their suitability for the appropriate satellite images using the neutrosophic weighted aggregated sum product assessment (WASPAS) method. The results obtained using the created methodology were verified with results acquired in an alternative way—using the edge detection algorithms for specific images. This methodology facilitates the selection of a proper edge detector for the chosen image content.This article belongs to the Collection Advanced Methodologies for Sustainability Assessment: Theory and Practic

    Taking Synchrony Seriously: A Perceptual-Level Model of Infant Synchrony Detection

    Get PDF
    Synchrony detection between different sensory and/or motor channels appears critically important for young infant learning and cognitive development. For example, empirical studies demonstrate that audio-visual synchrony aids in language acquisition. In this paper we compare these infant studies with a model of synchrony detection based on the Hershey and Movellan (2000) algorithm augmented with methods for quantitative synchrony estimation. Four infant-model comparisons are presented, using audio-visual stimuli of increasing complexity. While infants and the model showed learning or discrimination with each type of stimuli used, the model was most successful with stimuli comprised of one audio and one visual source, and also with two audio sources and a dynamic-face visual motion source. More difficult for the model were stimuli conditions with two motion sources, and more abstract visual dynamics—an oscilloscope instead of a face. Future research should model the developmental pathway of synchrony detection. Normal audio-visual synchrony detection in infants may be experience-dependent (e.g., Bergeson, et al., 2004)

    Beyond saliency: understanding convolutional neural networks from saliency prediction on layer-wise relevance propagation

    Get PDF
    Despite the tremendous achievements of deep convolutional neural networks (CNNs) in many computer vision tasks, understanding how they actually work remains a significant challenge. In this paper, we propose a novel two-step understanding method, namely Salient Relevance (SR) map, which aims to shed light on how deep CNNs recognize images and learn features from areas, referred to as attention areas, therein. Our proposed method starts out with a layer-wise relevance propagation (LRP) step which estimates a pixel-wise relevance map over the input image. Following, we construct a context-aware saliency map, SR map, from the LRP-generated map which predicts areas close to the foci of attention instead of isolated pixels that LRP reveals. In human visual system, information of regions is more important than of pixels in recognition. Consequently, our proposed approach closely simulates human recognition. Experimental results using the ILSVRC2012 validation dataset in conjunction with two well-established deep CNN models, AlexNet and VGG-16, clearly demonstrate that our proposed approach concisely identifies not only key pixels but also attention areas that contribute to the underlying neural network's comprehension of the given images. As such, our proposed SR map constitutes a convenient visual interface which unveils the visual attention of the network and reveals which type of objects the model has learned to recognize after training. The source code is available at https://github.com/Hey1Li/Salient-Relevance-Propagation.Comment: 35 pages, 15 figure

    Perception of Motion and Architectural Form: Computational Relationships between Optical Flow and Perspective

    Full text link
    Perceptual geometry refers to the interdisciplinary research whose objectives focuses on study of geometry from the perspective of visual perception, and in turn, applies such geometric findings to the ecological study of vision. Perceptual geometry attempts to answer fundamental questions in perception of form and representation of space through synthesis of cognitive and biological theories of visual perception with geometric theories of the physical world. Perception of form, space and motion are among fundamental problems in vision science. In cognitive and computational models of human perception, the theories for modeling motion are treated separately from models for perception of form.Comment: 10 pages, 13 figures, submitted and accepted in DoCEIS'2012 Conference: http://www.uninova.pt/doceis/doceis12/home/home.ph

    Luminance cues constrain chromatic blur discrimination in natural scene stimuli

    Get PDF
    Introducing blur into the color components of a natural scene has very little effect on its percept, whereas blur introduced into the luminance component is very noticeable. Here we quantify the dominance of luminance information in blur detection and examine a number of potential causes. We show that the interaction between chromatic and luminance information is not explained by reduced acuity or spatial resolution limitations for chromatic cues, the effective contrast of the luminance cue, or chromatic and achromatic statistical regularities in the images. Regardless of the quality of chromatic information, the visual system gives primacy to luminance signals when determining edge location. In natural viewing, luminance information appears to be specialized for detecting object boundaries while chromatic information may be used to determine surface properties
    corecore