3,255 research outputs found
GeoSay: A Geometric Saliency for Extracting Buildings in Remote Sensing Images
Automatic extraction of buildings in remote sensing images is an important
but challenging task and finds many applications in different fields such as
urban planning, navigation and so on. This paper addresses the problem of
buildings extraction in very high-spatial-resolution (VHSR) remote sensing (RS)
images, whose spatial resolution is often up to half meters and provides rich
information about buildings. Based on the observation that buildings in VHSR-RS
images are always more distinguishable in geometry than in texture or spectral
domain, this paper proposes a geometric building index (GBI) for accurate
building extraction, by computing the geometric saliency from VHSR-RS images.
More precisely, given an image, the geometric saliency is derived from a
mid-level geometric representations based on meaningful junctions that can
locally describe geometrical structures of images. The resulting GBI is finally
measured by integrating the derived geometric saliency of buildings.
Experiments on three public and commonly used datasets demonstrate that the
proposed GBI achieves the state-of-the-art performance and shows impressive
generalization capability. Additionally, GBI preserves both the exact position
and accurate shape of single buildings compared to existing methods
Motion clouds: model-based stimulus synthesis of natural-like random textures for the study of motion perception
Choosing an appropriate set of stimuli is essential to characterize the
response of a sensory system to a particular functional dimension, such as the
eye movement following the motion of a visual scene. Here, we describe a
framework to generate random texture movies with controlled information
content, i.e., Motion Clouds. These stimuli are defined using a generative
model that is based on controlled experimental parametrization. We show that
Motion Clouds correspond to dense mixing of localized moving gratings with
random positions. Their global envelope is similar to natural-like stimulation
with an approximate full-field translation corresponding to a retinal slip. We
describe the construction of these stimuli mathematically and propose an
open-source Python-based implementation. Examples of the use of this framework
are shown. We also propose extensions to other modalities such as color vision,
touch, and audition
Multi-scale Discriminant Saliency with Wavelet-based Hidden Markov Tree Modelling
The bottom-up saliency, an early stage of humans' visual attention, can be
considered as a binary classification problem between centre and surround
classes. Discriminant power of features for the classification is measured as
mutual information between distributions of image features and corresponding
classes . As the estimated discrepancy very much depends on considered scale
level, multi-scale structure and discriminant power are integrated by employing
discrete wavelet features and Hidden Markov Tree (HMT). With wavelet
coefficients and Hidden Markov Tree parameters, quad-tree like label structures
are constructed and utilized in maximum a posterior probability (MAP) of hidden
class variables at corresponding dyadic sub-squares. Then, a saliency value for
each square block at each scale level is computed with discriminant power
principle. Finally, across multiple scales is integrated the final saliency map
by an information maximization rule. Both standard quantitative tools such as
NSS, LCC, AUC and qualitative assessments are used for evaluating the proposed
multi-scale discriminant saliency (MDIS) method against the well-know
information based approach AIM on its released image collection with
eye-tracking data. Simulation results are presented and analysed to verify the
validity of MDIS as well as point out its limitation for further research
direction.Comment: arXiv admin note: substantial text overlap with arXiv:1301.396
- …