4,831 research outputs found
An infrastructure-free magnetic-based indoor positioning system with deep learning
POCI-01-0247-FEDER-033479Infrastructure-free Indoor Positioning Systems (IPS) are becoming popular due to their scalability and a wide range of applications. Such systems often rely on deployed Wi-Fi networks. However, their usability may be compromised, either due to scanning restrictions from recent Android versions or the proliferation of 5G technology. This raises the need for new infrastructure-free IPS independent of Wi-Fi networks. In this paper, we propose the use of magnetic field data for IPS, through Deep Neural Networks (DNN). Firstly, a dataset of human indoor trajectories was collected with different smartphones. Afterwards, a magnetic fingerprint was constructed and relevant features were extracted to train a DNN that returns a probability map of a user’s location. Finally, two postprocessing methods were applied to obtain the most probable location regions. We asserted the performance of our solution against a test dataset, which produced a Success Rate of around 80%. We believe that these results are competitive for an IPS based on a single sensing source. Moreover, the magnetic field can be used as an additional information layer to increase the robustness and redundancy of current multi-source IPS.publishersversionpublishe
Spatial and temporal background modelling of non-stationary visual scenes
PhDThe prevalence of electronic imaging systems in everyday life has become increasingly apparent
in recent years. Applications are to be found in medical scanning, automated manufacture, and
perhaps most significantly, surveillance. Metropolitan areas, shopping malls, and road traffic
management all employ and benefit from an unprecedented quantity of video cameras for monitoring
purposes. But the high cost and limited effectiveness of employing humans as the final
link in the monitoring chain has driven scientists to seek solutions based on machine vision techniques.
Whilst the field of machine vision has enjoyed consistent rapid development in the last
20 years, some of the most fundamental issues still remain to be solved in a satisfactory manner.
Central to a great many vision applications is the concept of segmentation, and in particular,
most practical systems perform background subtraction as one of the first stages of video
processing. This involves separation of ‘interesting foreground’ from the less informative but
persistent background. But the definition of what is ‘interesting’ is somewhat subjective, and
liable to be application specific. Furthermore, the background may be interpreted as including
the visual appearance of normal activity of any agents present in the scene, human or otherwise.
Thus a background model might be called upon to absorb lighting changes, moving trees and
foliage, or normal traffic flow and pedestrian activity, in order to effect what might be termed in
‘biologically-inspired’ vision as pre-attentive selection. This challenge is one of the Holy Grails
of the computer vision field, and consequently the subject has received considerable attention.
This thesis sets out to address some of the limitations of contemporary methods of background
segmentation by investigating methods of inducing local mutual support amongst pixels
in three starkly contrasting paradigms: (1) locality in the spatial domain, (2) locality in the shortterm
time domain, and (3) locality in the domain of cyclic repetition frequency.
Conventional per pixel models, such as those based on Gaussian Mixture Models, offer no
spatial support between adjacent pixels at all. At the other extreme, eigenspace models impose
a structure in which every image pixel bears the same relation to every other pixel. But Markov
Random Fields permit definition of arbitrary local cliques by construction of a suitable graph, and
3
are used here to facilitate a novel structure capable of exploiting probabilistic local cooccurrence
of adjacent Local Binary Patterns. The result is a method exhibiting strong sensitivity to multiple
learned local pattern hypotheses, whilst relying solely on monochrome image data.
Many background models enforce temporal consistency constraints on a pixel in attempt to
confirm background membership before being accepted as part of the model, and typically some
control over this process is exercised by a learning rate parameter. But in busy scenes, a true
background pixel may be visible for a relatively small fraction of the time and in a temporally
fragmented fashion, thus hindering such background acquisition. However, support in terms of
temporal locality may still be achieved by using Combinatorial Optimization to derive shortterm
background estimates which induce a similar consistency, but are considerably more robust
to disturbance. A novel technique is presented here in which the short-term estimates act as
‘pre-filtered’ data from which a far more compact eigen-background may be constructed.
Many scenes entail elements exhibiting repetitive periodic behaviour. Some road junctions
employing traffic signals are among these, yet little is to be found amongst the literature regarding
the explicit modelling of such periodic processes in a scene. Previous work focussing on gait
recognition has demonstrated approaches based on recurrence of self-similarity by which local
periodicity may be identified. The present work harnesses and extends this method in order
to characterize scenes displaying multiple distinct periodicities by building a spatio-temporal
model. The model may then be used to highlight abnormality in scene activity. Furthermore, a
Phase Locked Loop technique with a novel phase detector is detailed, enabling such a model to
maintain correct synchronization with scene activity in spite of noise and drift of periodicity.
This thesis contends that these three approaches are all manifestations of the same broad
underlying concept: local support in each of the space, time and frequency domains, and furthermore,
that the support can be harnessed practically, as will be demonstrated experimentally
Rethinking Closed-loop Training for Autonomous Driving
Recent advances in high-fidelity simulators have enabled closed-loop training
of autonomous driving agents, potentially solving the distribution shift in
training v.s. deployment and allowing training to be scaled both safely and
cheaply. However, there is a lack of understanding of how to build effective
training benchmarks for closed-loop training. In this work, we present the
first empirical study which analyzes the effects of different training
benchmark designs on the success of learning agents, such as how to design
traffic scenarios and scale training environments. Furthermore, we show that
many popular RL algorithms cannot achieve satisfactory performance in the
context of autonomous driving, as they lack long-term planning and take an
extremely long time to train. To address these issues, we propose trajectory
value learning (TRAVL), an RL-based driving agent that performs planning with
multistep look-ahead and exploits cheaply generated imagined data for efficient
learning. Our experiments show that TRAVL can learn much faster and produce
safer maneuvers compared to all the baselines. For more information, visit the
project website: https://waabi.ai/research/travlComment: ECCV 202
Attention and Anticipation in Fast Visual-Inertial Navigation
We study a Visual-Inertial Navigation (VIN) problem in which a robot needs to
estimate its state using an on-board camera and an inertial sensor, without any
prior knowledge of the external environment. We consider the case in which the
robot can allocate limited resources to VIN, due to tight computational
constraints. Therefore, we answer the following question: under limited
resources, what are the most relevant visual cues to maximize the performance
of visual-inertial navigation? Our approach has four key ingredients. First, it
is task-driven, in that the selection of the visual cues is guided by a metric
quantifying the VIN performance. Second, it exploits the notion of
anticipation, since it uses a simplified model for forward-simulation of robot
dynamics, predicting the utility of a set of visual cues over a future time
horizon. Third, it is efficient and easy to implement, since it leads to a
greedy algorithm for the selection of the most relevant visual cues. Fourth, it
provides formal performance guarantees: we leverage submodularity to prove that
the greedy selection cannot be far from the optimal (combinatorial) selection.
Simulations and real experiments on agile drones show that our approach ensures
state-of-the-art VIN performance while maintaining a lean processing time. In
the easy scenarios, our approach outperforms appearance-based feature selection
in terms of localization errors. In the most challenging scenarios, it enables
accurate visual-inertial navigation while appearance-based feature selection
fails to track robot's motion during aggressive maneuvers.Comment: 20 pages, 7 figures, 2 table
Versatile Multi-Contact Planning and Control for Legged Loco-Manipulation
Loco-manipulation planning skills are pivotal for expanding the utility of
robots in everyday environments. These skills can be assessed based on a
system's ability to coordinate complex holistic movements and multiple contact
interactions when solving different tasks. However, existing approaches have
been merely able to shape such behaviors with hand-crafted state machines,
densely engineered rewards, or pre-recorded expert demonstrations. Here, we
propose a minimally-guided framework that automatically discovers whole-body
trajectories jointly with contact schedules for solving general
loco-manipulation tasks in pre-modeled environments. The key insight is that
multi-modal problems of this nature can be formulated and treated within the
context of integrated Task and Motion Planning (TAMP). An effective bilevel
search strategy is achieved by incorporating domain-specific rules and
adequately combining the strengths of different planning techniques: trajectory
optimization and informed graph search coupled with sampling-based planning. We
showcase emergent behaviors for a quadrupedal mobile manipulator exploiting
both prehensile and non-prehensile interactions to perform real-world tasks
such as opening/closing heavy dishwashers and traversing spring-loaded doors.
These behaviors are also deployed on the real system using a two-layer
whole-body tracking controller
INFN What Next: Ultra-relativistic Heavy-Ion Collisions
This document was prepared by the community that is active in Italy, within
INFN (Istituto Nazionale di Fisica Nucleare), in the field of
ultra-relativistic heavy-ion collisions. The experimental study of the phase
diagram of strongly-interacting matter and of the Quark-Gluon Plasma (QGP)
deconfined state will proceed, in the next 10-15 years, along two directions:
the high-energy regime at RHIC and at the LHC, and the low-energy regime at
FAIR, NICA, SPS and RHIC. The Italian community is strongly involved in the
present and future programme of the ALICE experiment, the upgrade of which will
open, in the 2020s, a new phase of high-precision characterisation of the QGP
properties at the LHC. As a complement of this main activity, there is a
growing interest in a possible future experiment at the SPS, which would target
the search for the onset of deconfinement using dimuon measurements. On a
longer timescale, the community looks with interest at the ongoing studies and
discussions on a possible fixed-target programme using the LHC ion beams and on
the Future Circular Collider.Comment: 99 pages, 56 figure
- …