5,952 research outputs found
Muon Track Reconstruction and Data Selection Techniques in AMANDA
The Antarctic Muon And Neutrino Detector Array (AMANDA) is a high-energy
neutrino telescope operating at the geographic South Pole. It is a lattice of
photo-multiplier tubes buried deep in the polar ice between 1500m and 2000m.
The primary goal of this detector is to discover astrophysical sources of high
energy neutrinos. A high-energy muon neutrino coming through the earth from the
Northern Hemisphere can be identified by the secondary muon moving upward
through the detector. The muon tracks are reconstructed with a maximum
likelihood method. It models the arrival times and amplitudes of Cherenkov
photons registered by the photo-multipliers. This paper describes the different
methods of reconstruction, which have been successfully implemented within
AMANDA. Strategies for optimizing the reconstruction performance and rejecting
background are presented. For a typical analysis procedure the direction of
tracks are reconstructed with about 2 degree accuracy.Comment: 40 pages, 16 Postscript figures, uses elsart.st
Multisource Self-calibration for Sensor Arrays
Calibration of a sensor array is more involved if the antennas have direction
dependent gains and multiple calibrator sources are simultaneously present. We
study this case for a sensor array with arbitrary geometry but identical
elements, i.e. elements with the same direction dependent gain pattern. A
weighted alternating least squares (WALS) algorithm is derived that iteratively
solves for the direction independent complex gains of the array elements, their
noise powers and their gains in the direction of the calibrator sources. An
extension of the problem is the case where the apparent calibrator source
locations are unknown, e.g., due to refractive propagation paths. For this
case, the WALS method is supplemented with weighted subspace fitting (WSF)
direction finding techniques. Using Monte Carlo simulations we demonstrate that
both methods are asymptotically statistically efficient and converge within two
iterations even in cases of low SNR.Comment: 11 pages, 8 figure
Laboratory and Field Experimental Study of Underwater Inflatable Co-prime Sonar Array (UICSA)
This paper discusses the design and initial testing of a novel hydrophone array system dubbed the Underwater Inflatable Co-prime Sonar Array (UICSA). The UICSA will be a crucial component of an underwater deployable sensing network that can be rapidly deployed using compact autonomous underwater vehicles (AUVs). The UICSA initially is packed in a compact container to fit the payload space of an AUV. After deployment, the UICSA expands to its predetermined full length to acquire sensing data for source localization. More specifically, the mechanical compression of the UICSA is achieved through a non-rigid array support structure, which consists of flexible inflatable segments between adjoining hydrophones that are folded in order to package the UICSA for deployment. The system exploits compression in hydrophone layouts by utilizing a sparse array configuration, namely the co-prime array since it requires fewer hydrophones than a uniform linear array of the same length to estimate a given number of sources. With two-way compression, the storage, handling, and transportation of the compactly designed UICSA is convenient, particularly for the AUVs with limited payload space. The deployment concept and process are discussed, as well as the various UICSA designs of different support structures are described. A comparison of the various mechanical designs is presented and a novel hybrid-based expansion prototype is documented in detail. Laboratory study results of the UICSA prototype are presented that include water-swollen material tests in a pressurized environment and water tank validation of the inflation process. The UICSA prototype also has been deployed in the Harbor Branch channel to validate the performance, the related field test details and source localization results
Towards Odor-Sensitive Mobile Robots
J. Monroy, J. Gonzalez-Jimenez, "Towards Odor-Sensitive Mobile Robots", Electronic Nose Technologies and Advances in Machine Olfaction, IGI Global, pp. 244--263, 2018, doi:10.4018/978-1-5225-3862-2.ch012
Versión preprint, con permiso del editorOut of all the components of a mobile robot, its sensorial system is undoubtedly among the most critical
ones when operating in real environments. Until now, these sensorial systems mostly relied on range
sensors (laser scanner, sonar, active triangulation) and cameras. While electronic noses have barely
been employed, they can provide a complementary sensory information, vital for some applications, as
with humans. This chapter analyzes the motivation of providing a robot with gas-sensing capabilities
and also reviews some of the hurdles that are preventing smell from achieving the importance of other
sensing modalities in robotics. The achievements made so far are reviewed to illustrate the current status
on the three main fields within robotics olfaction: the classification of volatile substances, the spatial
estimation of the gas dispersion from sparse measurements, and the localization of the gas source within
a known environment
Plenoptic Signal Processing for Robust Vision in Field Robotics
This thesis proposes the use of plenoptic cameras for improving the robustness and simplicity of machine vision in field robotics applications. Dust, rain, fog, snow, murky water and insufficient light can cause even the most sophisticated vision systems to fail. Plenoptic cameras offer an appealing alternative to conventional imagery by gathering significantly more light over a wider depth of field, and capturing a rich 4D light field structure that encodes textural and geometric information. The key contributions of this work lie in exploring the properties of plenoptic signals and developing algorithms for exploiting them. It lays the groundwork for the deployment of plenoptic cameras in field robotics by establishing a decoding, calibration and rectification scheme appropriate to compact, lenslet-based devices. Next, the frequency-domain shape of plenoptic signals is elaborated and exploited by constructing a filter which focuses over a wide depth of field rather than at a single depth. This filter is shown to reject noise, improving contrast in low light and through attenuating media, while mitigating occluders such as snow, rain and underwater particulate matter. Next, a closed-form generalization of optical flow is presented which directly estimates camera motion from first-order derivatives. An elegant adaptation of this "plenoptic flow" to lenslet-based imagery is demonstrated, as well as a simple, additive method for rendering novel views. Finally, the isolation of dynamic elements from a static background is considered, a task complicated by the non-uniform apparent motion caused by a mobile camera. Two elegant closed-form solutions are presented dealing with monocular time-series and light field image pairs. This work emphasizes non-iterative, noise-tolerant, closed-form, linear methods with predictable and constant runtimes, making them suitable for real-time embedded implementation in field robotics applications
Plenoptic Signal Processing for Robust Vision in Field Robotics
This thesis proposes the use of plenoptic cameras for improving the robustness and simplicity of machine vision in field robotics applications. Dust, rain, fog, snow, murky water and insufficient light can cause even the most sophisticated vision systems to fail. Plenoptic cameras offer an appealing alternative to conventional imagery by gathering significantly more light over a wider depth of field, and capturing a rich 4D light field structure that encodes textural and geometric information. The key contributions of this work lie in exploring the properties of plenoptic signals and developing algorithms for exploiting them. It lays the groundwork for the deployment of plenoptic cameras in field robotics by establishing a decoding, calibration and rectification scheme appropriate to compact, lenslet-based devices. Next, the frequency-domain shape of plenoptic signals is elaborated and exploited by constructing a filter which focuses over a wide depth of field rather than at a single depth. This filter is shown to reject noise, improving contrast in low light and through attenuating media, while mitigating occluders such as snow, rain and underwater particulate matter. Next, a closed-form generalization of optical flow is presented which directly estimates camera motion from first-order derivatives. An elegant adaptation of this "plenoptic flow" to lenslet-based imagery is demonstrated, as well as a simple, additive method for rendering novel views. Finally, the isolation of dynamic elements from a static background is considered, a task complicated by the non-uniform apparent motion caused by a mobile camera. Two elegant closed-form solutions are presented dealing with monocular time-series and light field image pairs. This work emphasizes non-iterative, noise-tolerant, closed-form, linear methods with predictable and constant runtimes, making them suitable for real-time embedded implementation in field robotics applications
Algorithms for propagation-aware underwater ranging and localization
Mención Internacional en el título de doctorWhile oceans occupy most of our planet, their exploration and conservation are one of
the crucial research problems of modern time. Underwater localization stands among the
key issues on the way to the proper inspection and monitoring of this significant part of our
world. In this thesis, we investigate and tackle different challenges related to underwater
ranging and localization. In particular, we focus on algorithms that consider underwater
acoustic channel properties. This group of algorithms utilizes additional information
about the environment and its impact on acoustic signal propagation, in order to improve
the accuracy of location estimates, or to achieve a reduced complexity, or a reduced
amount of resources (e.g., anchor nodes) compared to traditional algorithms.
First, we tackle the problem of passive range estimation using the differences in the
times of arrival of multipath replicas of a transmitted acoustic signal. This is a costand
energy- effective algorithm that can be used for the localization of autonomous
underwater vehicles (AUVs), and utilizes information about signal propagation. We study
the accuracy of this method in the simplified case of constant sound speed profile (SSP)
and compare it to a more realistic case with various non-constant SSP. We also propose
an auxiliary quantity called effective sound speed. This quantity, when modeling acoustic
propagation via ray models, takes into account the difference between rectilinear and
non-rectilinear sound ray paths. According to our evaluation, this offers improved range
estimation results with respect to standard algorithms that consider the actual value of
the speed of sound.
We then propose an algorithm suitable for the non-invasive tracking of AUVs or
vocalizing marine animals, using only a single receiver. This algorithm evaluates the
underwater acoustic channel impulse response differences induced by a diverse sea
bottom profile, and proposes a computationally- and energy-efficient solution for passive
localization.
Finally, we propose another algorithm to solve the issue of 3D acoustic localization
and tracking of marine fauna. To reach the expected degree of accuracy, more sensors
are often required than are available in typical commercial off-the-shelf (COTS) phased
arrays found, e.g., in ultra short baseline (USBL) systems. Direct combination of multiple
COTS arrays may be constrained by array body elements, and lead to breaking the optimal array element spacing, or the desired array layout. Thus, the application of
state-of-the-art direction of arrival (DoA) estimation algorithms may not be possible. We
propose a solution for passive 3D localization and tracking using a wideband acoustic
array of arbitrary shape, and validate the algorithm in multiple experiments, involving
both active and passive targets.Part of the research in this thesis has been supported by the EU H2020 program under
project SYMBIOSIS (G.A. no. 773753).This work has been supported by IMDEA Networks InstitutePrograma de Doctorado en Ingeniería Telemática por la Universidad Carlos III de MadridPresidente: Paul Daniel Mitchell.- Secretario: Antonio Fernández Anta.- Vocal: Santiago Zazo Bell
Calibration and image formation in the ABACUS sonar system
Includes bibliographical references (leaves 138-139)
- …