1,098 research outputs found
Recognizing point clouds using conditional random fields
Detecting objects in cluttered scenes is a necessary step for many robotic tasks and facilitates the interaction of the robot with its environment. Because of the availability of efficient 3D sensing devices as the Kinect, methods for the recognition of objects in 3D point clouds have gained importance during the last years. In this paper, we propose a new supervised learning approach for the recognition of objects from 3D point clouds using Conditional Random Fields, a type of discriminative, undirected probabilistic graphical model. The various features and contextual relations of the objects are described by the potential functions in the graph. Our method allows for learning and inference from unorganized point clouds of arbitrary sizes and shows significant benefit in terms of computational speed during prediction when compared to a state-of-the-art approach based on constrained optimization.Peer ReviewedPostprint (author’s final draft
Online Context-based Object Recognition for Mobile Robots
This work proposes a robotic object recognition
system that takes advantage of the contextual information latent
in human-like environments in an online fashion. To fully leverage
context, it is needed perceptual information from (at least) a
portion of the scene containing the objects of interest, which could
not be entirely covered by just an one-shot sensor observation.
Information from a larger portion of the scenario could still
be considered by progressively registering observations, but this
approach experiences difficulties under some circumstances, e.g.
limited and heavily demanded computational resources, dynamic
environments, etc. Instead of this, the proposed recognition
system relies on an anchoring process for the fast registration
and propagation of objects’ features and locations beyond the
current sensor frustum. In this way, the system builds a graphbased
world model containing the objects in the scenario (both
in the current and previously perceived shots), which is exploited
by a Probabilistic Graphical Model (PGM) in order to leverage
contextual information during recognition. We also propose a
novel way to include the outcome of local object recognition
methods in the PGM, which results in a decrease in the usually
high CRF learning complexity. A demonstration of our proposal
has been conducted employing a dataset captured by a mobile
robot from restaurant-like settings, showing promising results.Universidad de Málaga. Campus de Excelencia Internacional AndalucÃa Tech
Space-Time Hierarchical-Graph Based Cooperative Localization in Wireless Sensor Networks
It has been shown that cooperative localization is capable of improving both
the positioning accuracy and coverage in scenarios where the global positioning
system (GPS) has a poor performance. However, due to its potentially excessive
computational complexity, at the time of writing the application of cooperative
localization remains limited in practice. In this paper, we address the
efficient cooperative positioning problem in wireless sensor networks. A
space-time hierarchical-graph based scheme exhibiting fast convergence is
proposed for localizing the agent nodes. In contrast to conventional methods,
agent nodes are divided into different layers with the aid of the space-time
hierarchical-model and their positions are estimated gradually. In particular,
an information propagation rule is conceived upon considering the quality of
positional information. According to the rule, the information always
propagates from the upper layers to a certain lower layer and the message
passing process is further optimized at each layer. Hence, the potential error
propagation can be mitigated. Additionally, both position estimation and
position broadcasting are carried out by the sensor nodes. Furthermore, a
sensor activation mechanism is conceived, which is capable of significantly
reducing both the energy consumption and the network traffic overhead incurred
by the localization process. The analytical and numerical results provided
demonstrate the superiority of our space-time hierarchical-graph based
cooperative localization scheme over the benchmarking schemes considered.Comment: 14 pages, 15 figures, 4 tables, accepted to appear on IEEE
Transactions on Signal Processing, Sept. 201
Rich probabilistic models for semantic labeling
Das Ziel dieser Monographie ist es die Methoden und Anwendungen des semantischen Labelings zu erforschen. Unsere Beiträge zu diesem sich rasch entwickelten Thema sind bestimmte Aspekte der Modellierung und der Inferenz in probabilistischen Modellen und ihre Anwendungen in den interdisziplinären Bereichen der Computer Vision sowie medizinischer Bildverarbeitung und Fernerkundung
Development of a probabilistic perception system for camera-lidar sensor fusion
La estimación de profundidad usando diferentes sensores es uno de los desafÃos clave para dotar a las máquinas autónomas de sólidas capacidades de percepción robótica. Ha habido un avance sobresaliente en el desarrollo de técnicas de estimación de profundidad unimodales basadas en cámaras monoculares, debido a su alta resolución o sensores LiDAR, debido a los datos geométricos precisos que proporcionan. Sin embargo, cada uno de ellos presenta inconvenientes inherentes, como la alta sensibilidad a los cambios en las condiciones de iluminación en el caso delas cámaras y la resolución limitada de los sensores LiDAR. La fusión de sensores se puede utilizar para combinar los méritos y compensar las desventajas de estos dos tipos de sensores. Sin embargo, los métodos de fusión actuales funcionan a un alto nivel. Procesan los flujos de datos de los sensores de forma independiente y combinan las estimaciones de alto nivel obtenidas para cada sensor. En este proyecto, abordamos el problema en un nivel bajo, fusionando los flujos de sensores sin procesar, obteniendo asà estimaciones de profundidad que son densas y precisas, y pueden usarse como una fuente de datos multimodal unificada para problemas de estimación de nivel superior. Este trabajo propone un modelo de campo aleatorio condicional (CRF) con múltiples potenciales de geometrÃa y apariencia que representa a la perfección el problema de estimar mapas de profundidad densos a partir de datos de cámara y LiDAR. El modelo se puede optimizar de manera eficiente utilizando el algoritmo Conjúgate Gradient Squared (CGS). El método propuesto se evalúa y compara utilizando el conjunto de datos proporcionado por KITTI Datset. Adicionalmente, se evalúa cualitativamente el modelo, usando datos adquiridos por el autor de esté trabajoMulti-modal depth estimation is one of the key challenges for endowing autonomous
machines with robust robotic perception capabilities. There has been an outstanding
advance in the development of uni-modal depth estimation techniques based
on either monocular cameras, because of their rich resolution or LiDAR sensors due
to the precise geometric data they provide. However, each of them suffers from some
inherent drawbacks like high sensitivity to changes in illumination conditions in
the case of cameras and limited resolution for the LiDARs. Sensor fusion can be
used to combine the merits and compensate the downsides of these two kinds of
sensors. Nevertheless, current fusion methods work at a high level. They processes
sensor data streams independently and combine the high level estimates obtained
for each sensor. In this thesis, I tackle the problem at a low level, fusing the raw
sensor streams, thus obtaining depth estimates which are both dense and precise,
and can be used as a unified multi-modal data source for higher level estimation
problems.
This work proposes a Conditional Random Field (CRF) model with multiple geometry
and appearance potentials that seamlessly represents the problem of estimating
dense depth maps from camera and LiDAR data. The model can be optimized
efficiently using the Conjugate Gradient Squared (CGS) algorithm. The proposed
method was evaluated and compared with the state-of-the-art using the commonly
used KITTI benchmark dataset. In addition, the model is qualitatively evaluated using
data acquired by the author of this work.MaestrÃaMagÃster en IngenierÃa de Desarrollo de Producto
Target Tracking in Confined Environments with Uncertain Sensor Positions
To ensure safety in confined environments such as mines or subway tunnels, a
(wireless) sensor network can be deployed to monitor various environmental
conditions. One of its most important applications is to track personnel,
mobile equipment and vehicles. However, the state-of-the-art algorithms assume
that the positions of the sensors are perfectly known, which is not necessarily
true due to imprecise placement and/or dropping of sensors. Therefore, we
propose an automatic approach for simultaneous refinement of sensors' positions
and target tracking. We divide the considered area in a finite number of cells,
define dynamic and measurement models, and apply a discrete variant of belief
propagation which can efficiently solve this high-dimensional problem, and
handle all non-Gaussian uncertainties expected in this kind of environments.
Finally, we use ray-tracing simulation to generate an artificial mine-like
environment and generate synthetic measurement data. According to our extensive
simulation study, the proposed approach performs significantly better than
standard Bayesian target tracking and localization algorithms, and provides
robustness against outliers.Comment: IEEE Transactions on Vehicular Technology, 201
Belief Consensus Algorithms for Fast Distributed Target Tracking in Wireless Sensor Networks
In distributed target tracking for wireless sensor networks, agreement on the
target state can be achieved by the construction and maintenance of a
communication path, in order to exchange information regarding local likelihood
functions. Such an approach lacks robustness to failures and is not easily
applicable to ad-hoc networks. To address this, several methods have been
proposed that allow agreement on the global likelihood through fully
distributed belief consensus (BC) algorithms, operating on local likelihoods in
distributed particle filtering (DPF). However, a unified comparison of the
convergence speed and communication cost has not been performed. In this paper,
we provide such a comparison and propose a novel BC algorithm based on belief
propagation (BP). According to our study, DPF based on metropolis belief
consensus (MBC) is the fastest in loopy graphs, while DPF based on BP consensus
is the fastest in tree graphs. Moreover, we found that BC-based DPF methods
have lower communication overhead than data flooding when the network is
sufficiently sparse
- …