51 research outputs found
Multi-Modal Trip Hazard Affordance Detection On Construction Sites
Trip hazards are a significant contributor to accidents on construction and
manufacturing sites, where over a third of Australian workplace injuries occur
[1]. Current safety inspections are labour intensive and limited by human
fallibility,making automation of trip hazard detection appealing from both a
safety and economic perspective. Trip hazards present an interesting challenge
to modern learning techniques because they are defined as much by affordance as
by object type; for example wires on a table are not a trip hazard, but can be
if lying on the ground. To address these challenges, we conduct a comprehensive
investigation into the performance characteristics of 11 different colour and
depth fusion approaches, including 4 fusion and one non fusion approach; using
colour and two types of depth images. Trained and tested on over 600 labelled
trip hazards over 4 floors and 2000m in an active construction
site,this approach was able to differentiate between identical objects in
different physical configurations (see Figure 1). Outperforming a colour-only
detector, our multi-modal trip detector fuses colour and depth information to
achieve a 4% absolute improvement in F1-score. These investigative results and
the extensive publicly available dataset moves us one step closer to assistive
or fully automated safety inspection systems on construction sites.Comment: 9 Pages, 12 Figures, 2 Tables, Accepted to Robotics and Automation
Letters (RA-L
Multisource and Multitemporal Data Fusion in Remote Sensing
The sharp and recent increase in the availability of data captured by
different sensors combined with their considerably heterogeneous natures poses
a serious challenge for the effective and efficient processing of remotely
sensed data. Such an increase in remote sensing and ancillary datasets,
however, opens up the possibility of utilizing multimodal datasets in a joint
manner to further improve the performance of the processing approaches with
respect to the application at hand. Multisource data fusion has, therefore,
received enormous attention from researchers worldwide for a wide variety of
applications. Moreover, thanks to the revisit capability of several spaceborne
sensors, the integration of the temporal information with the spatial and/or
spectral/backscattering information of the remotely sensed data is possible and
helps to move from a representation of 2D/3D data to 4D data structures, where
the time variable adds new information as well as challenges for the
information extraction algorithms. There are a huge number of research works
dedicated to multisource and multitemporal data fusion, but the methods for the
fusion of different modalities have expanded in different paths according to
each research community. This paper brings together the advances of multisource
and multitemporal data fusion approaches with respect to different research
communities and provides a thorough and discipline-specific starting point for
researchers at different levels (i.e., students, researchers, and senior
researchers) willing to conduct novel investigations on this challenging topic
by supplying sufficient detail and references
Multi-source Pseudo-label Learning of Semantic Segmentation for the Scene Recognition of Agricultural Mobile Robots
This paper describes a novel method of training a semantic segmentation model
for environment recognition of agricultural mobile robots by unsupervised
domain adaptation exploiting publicly available datasets of outdoor scenes that
are different from our target environments i.e., greenhouses. In conventional
semantic segmentation methods, the labels are given by manual annotation, which
is a tedious and time-consuming task. A method to work around the necessity of
the manual annotation is unsupervised domain adaptation (UDA) that transfer
knowledge from labeled source datasets to unlabeled target datasets. Most of
the UDA methods of semantic segmentation are validated by tasks of adaptation
from non-photorealistic synthetic images of urban scenes to real ones. However,
the effectiveness of the methods is not well studied in the case of adaptation
to other types of environments, such as greenhouses. In addition, it is not
always possible to prepare appropriate source datasets for such environments.
In this paper, we adopt an existing training method of UDA to a task of
training a model for greenhouse images. We propose to use multiple publicly
available datasets of outdoor images as source datasets, and also propose a
simple yet effective method of generating pseudo-labels by transferring
knowledge from the source datasets that have different appearance and a label
set from the target datasets. We demonstrate in experiments that by combining
our proposed method of pseudo-label generation with the existing training
method, the performance was improved by up to 14.3% of mIoU compared to the
best score of the single-source training.Comment: 10 pages, 7 figures, submitted to Machine Vision And Application
- …