17,725 research outputs found
Ground-based hyperspectral analysis of the urban nightscape
Airborne hyperspectral cameras provide the basic information to estimate the energy wasted skywards by outdoor lighting systems, as well as to locate and identify their sources. However, a complete characterization of the urban light pollution levels also requires evaluating these effects from the city dwellers standpoint, e.g. the energy waste associated to the excessive illuminance on walls and pavements, light trespass, or the luminance distributions causing potential glare, to mention but a few. On the other hand, the spectral irradiance at the entrance of the human eye is the primary input to evaluate the possible health effects associated with the exposure to artificial light at night, according to the more recent models available in the literature. In this work we demonstrate the possibility of using a hyperspectral imager (routinely used in airborne campaigns) to measure the ground-level spectral radiance of the urban nightscape and to retrieve several magnitudes of interest for light pollution studies. We also present the preliminary results from a field campaign carried out in the downtown of Barcelona.Postprint (author's final draft
LCrowdV: Generating Labeled Videos for Simulation-based Crowd Behavior Learning
We present a novel procedural framework to generate an arbitrary number of
labeled crowd videos (LCrowdV). The resulting crowd video datasets are used to
design accurate algorithms or training models for crowded scene understanding.
Our overall approach is composed of two components: a procedural simulation
framework for generating crowd movements and behaviors, and a procedural
rendering framework to generate different videos or images. Each video or image
is automatically labeled based on the environment, number of pedestrians,
density, behavior, flow, lighting conditions, viewpoint, noise, etc.
Furthermore, we can increase the realism by combining synthetically-generated
behaviors with real-world background videos. We demonstrate the benefits of
LCrowdV over prior lableled crowd datasets by improving the accuracy of
pedestrian detection and crowd behavior classification algorithms. LCrowdV
would be released on the WWW
Modeling Camera Effects to Improve Visual Learning from Synthetic Data
Recent work has focused on generating synthetic imagery to increase the size
and variability of training data for learning visual tasks in urban scenes.
This includes increasing the occurrence of occlusions or varying environmental
and weather effects. However, few have addressed modeling variation in the
sensor domain. Sensor effects can degrade real images, limiting
generalizability of network performance on visual tasks trained on synthetic
data and tested in real environments. This paper proposes an efficient,
automatic, physically-based augmentation pipeline to vary sensor effects
--chromatic aberration, blur, exposure, noise, and color cast-- for synthetic
imagery. In particular, this paper illustrates that augmenting synthetic
training datasets with the proposed pipeline reduces the domain gap between
synthetic and real domains for the task of object detection in urban driving
scenes
Fair comparison of skin detection approaches on publicly available datasets
Skin detection is the process of discriminating skin and non-skin regions in
a digital image and it is widely used in several applications ranging from hand
gesture analysis to track body parts and face detection. Skin detection is a
challenging problem which has drawn extensive attention from the research
community, nevertheless a fair comparison among approaches is very difficult
due to the lack of a common benchmark and a unified testing protocol. In this
work, we investigate the most recent researches in this field and we propose a
fair comparison among approaches using several different datasets. The major
contributions of this work are an exhaustive literature review of skin color
detection approaches, a framework to evaluate and combine different skin
detector approaches, whose source code is made freely available for future
research, and an extensive experimental comparison among several recent methods
which have also been used to define an ensemble that works well in many
different problems. Experiments are carried out in 10 different datasets
including more than 10000 labelled images: experimental results confirm that
the best method here proposed obtains a very good performance with respect to
other stand-alone approaches, without requiring ad hoc parameter tuning. A
MATLAB version of the framework for testing and of the methods proposed in this
paper will be freely available from https://github.com/LorisNann
‘Do you see what I see?’ Medical imaging: the interpretation of visual information
Röntgen's discovery of x-rays in 1895, gave to medicine the extraordinary benefit of being able to see inside the living body without surgery. Over time, technology has added to the sophistication of imaging processes in medicine and we now have a wide range of techniques at our disposal for the investigation and early detection of disease. But radiology deals with visual information; and like any information this requires interpretation. It is a practical field and medical images are used to make inferences about the state of peoples' health. These inferences are subject to the same variability and error as any decision-making process and so the criteria for the success of medical imaging are based not entirely on the images themselves but on the performance of the decision-makers. Research in the accuracy of medical imaging must draw on techniques from a wide range of disciplines including physics, psychology, computing, neuroscience and medicine in attempting to better understand the processes involved in visual decision-making in this context and to minimise diagnostic error
- …