44 research outputs found
MEMS Accelerometers
Micro-electro-mechanical system (MEMS) devices are widely used for inertia, pressure, and ultrasound sensing applications. Research on integrated MEMS technology has undergone extensive development driven by the requirements of a compact footprint, low cost, and increased functionality. Accelerometers are among the most widely used sensors implemented in MEMS technology. MEMS accelerometers are showing a growing presence in almost all industries ranging from automotive to medical. A traditional MEMS accelerometer employs a proof mass suspended to springs, which displaces in response to an external acceleration. A single proof mass can be used for one- or multi-axis sensing. A variety of transduction mechanisms have been used to detect the displacement. They include capacitive, piezoelectric, thermal, tunneling, and optical mechanisms. Capacitive accelerometers are widely used due to their DC measurement interface, thermal stability, reliability, and low cost. However, they are sensitive to electromagnetic field interferences and have poor performance for high-end applications (e.g., precise attitude control for the satellite). Over the past three decades, steady progress has been made in the area of optical accelerometers for high-performance and high-sensitivity applications but several challenges are still to be tackled by researchers and engineers to fully realize opto-mechanical accelerometers, such as chip-scale integration, scaling, low bandwidth, etc
Sistemas de posicionamento baseados em comunicação por luz para ambientes interiores
The demand for highly precise indoor positioning systems (IPSs) is growing
rapidly due to its potential in the increasingly popular techniques of the
Internet of Things, smart mobile devices, and artificial intelligence. IPS
becomes a promising research domain that is getting wide attention due to its
benefits in several working scenarios, such as, industries, indoor public
locations, and autonomous navigation. Moreover, IPS has a prominent
contribution in day-to-day activities in organizations such as health care
centers, airports, shopping malls, manufacturing, underground locations, etc.,
for safe operating environments. In indoor environments, both radio frequency
(RF) and optical wireless communication (OWC) based technologies could be
adopted for localization. Although the RF-based global positioning system,
such as, Global positioning system offers higher penetration rates with
reduced accuracy (i.e., in the range of a few meters), it does not work well in
indoor environments (and not at all in certain cases such as tunnels, mines,
etc.) due to the very weak signal and no direct access to the satellites. On the
other hand, the light-based system known as a visible light positioning (VLP)
system, as part of the OWC systems, uses the pre-existing light-emitting
diodes (LEDs)-based lighting infrastructure, could be used at low cost and
high accuracy compared with the RF-based systems. VLP is an emerging
technology promising high accuracy, high security, low deployment cost,
shorter time response, and low relative complexity when compared with RFbased
positioning.
However, in indoor VLP systems, there are some concerns such as,
multipath reflection, transmitter tilting, transmitter’s position, and orientation
uncertainty, human shadowing/blocking, and noise causing the increase in
the positioning error, thereby reducing the positioning accuracy of the system.
Therefore, it is imperative to capture the characteristics of different VLP
channel and properly model them for the dual purpose of illumination and
localization. In this thesis, firstly, the impact of transmitter tilting angles and
multipath reflections are studied and for the first time, it is demonstrated that
tilting the transmitter can be beneficial in VLP systems considering both line of
sight (LOS) and non-line of sight transmission paths. With the transmitters
oriented towards the center of the receiving plane, the received power level is
maximized due to the LOS components. It is also shown that the proposed
scheme offers a significant accuracy improvement of up to ~66% compared
with a typical non-tilted transmitter VLP. The effect of tilting the transmitter on
the lighting uniformity is also investigated and results proved that the
uniformity achieved complies with the European Standard EN 12464-1.
After that, the impact of transmitter position and orientation uncertainty on
the accuracy of the VLP system based on the received signal strength (RSS)
is investigated. Simulation results show that the transmitter uncertainties have
a severe impact on the positioning error, which can be leveraged through the
usage of more transmitters. Concerning a smaller transmitter’s position
epochs, and the size of the training set. It is shown that,
the ANN with Bayesian regularization outperforms the traditional RSS
technique using the non-linear least square estimation for all values of signal
to noise ratio.
Furthermore, a novel indoor VLP system is proposed based on support
vector machines and polynomial regression considering two different
multipath environments of an empty room and a furnished room. The results
show that, in an empty room, the positioning accuracy improvement for the
positioning error of 2.5 cm are 36.1, 58.3, and 72.2 % for three different
scenarios according to the regions’ distribution in the room. For the furnished
room, a positioning relative accuracy improvement of 214, 170, and 100 % is
observed for positioning error of 0.1, 0.2, and 0.3 m, respectively. Ultimately,
an indoor VLP system based on convolutional neural networks (CNN) is
proposed and demonstrated experimentally in which LEDs are used as
transmitters and a rolling shutter camera is used as receiver. A detection
algorithm named single shot detector (SSD) is used which relies on CNN (i.e.,
MobileNet or ResNet) for classification as well as position estimation of each
LED in the image. The system is validated using a real-world size test setup
containing eight LED luminaries. The obtained results show that the maximum
average root mean square positioning error achieved is 4.67 and 5.27 cm with
SSD MobileNet and SSD ResNet models, respectively. The validation results
show that the system can process 67 images per second, allowing real-time
positioning.A procura por sistemas de posicionamento interior (IPSs) de alta precisão tem
crescido rapidamente devido ao seu interesse nas técnicas cada vez mais
populares da Internet das Coisas, dispositivos móveis inteligentes e
inteligência artificial. O IPS tornou-se um domínio de pesquisa promissor que
tem atraído grande atenção devido aos seus benefícios em vários cenários de
trabalho, como indústrias, locais públicos e navegação autónoma. Além disso,
o IPS tem uma contribuição destacada no dia a dia de organizações, como,
centros de saúde, aeroportos, supermercados, fábricas, locais subterrâneos,
etc. As tecnologias baseadas em radiofrequência (RF) e comunicação óptica
sem fio (OWC) podem ser adotadas para localização em ambientes interiores.
Embora o sistema de posicionamento global (GPS) baseado em RF ofereça
taxas de penetração mais altas com precisão reduzida (ou seja, na faixa de
alguns metros), não funciona bem em ambientes interiores (e não funciona
bem em certos casos como túneis, minas, etc.) devido ao sinal muito fraco e
falta de acesso direto aos satélites. Por outro lado, o sistema baseado em luz
conhecido como sistema de posicionamento de luz visível (VLP), como parte
dos sistemas OWC, usa a infraestrutura de iluminação baseada em díodos
emissores de luz (LEDs) pré-existentes, é um sistemas de baixo custo e alta
precisão quando comprado com os sistemas baseados em RF. O VLP é uma
tecnologia emergente que promete alta precisão, alta segurança, baixo custo
de implantação, menor tempo de resposta e baixa complexidade relativa
quando comparado ao posicionamento baseado em RF.
No entanto, os sistemas VLP interiores, exibem algumas limitações, como, a
reflexão multicaminho, inclinação do transmissor, posição do transmissor e
incerteza de orientação, sombra/bloqueio humano e ruído, que têm como
consequência o aumento do erro de posicionamento, e consequente redução
da precisão do sistema. Portanto, é imperativo estudar as características dos
diferentes canais VLP e modelá-los adequadamente para o duplo propósito de
iluminação e localização. Esta tesa aborda, primeiramente, o impacto dos
ângulos de inclinação do transmissor e reflexões multipercurso no
desempenho do sistema de posicionamento. Demonstra-se que a inclinação
do transmissor pode ser benéfica em sistemas VLP considerando tanto a linha
de vista (LOS) como as reflexões. Com os transmissores orientados para o
centro do plano recetor, o nível de potência recebido é maximizado devido aos
componentes LOS. Também é mostrado que o esquema proposto oferece
uma melhoria significativa de precisão de até ~66% em comparação com um
sistema VLP de transmissor não inclinado típico. O efeito da inclinação do
transmissor na uniformidade da iluminação também é investigado e os
resultados comprovam que a uniformidade alcançada está de acordo com a
Norma Europeia EN 12464-1.
O impacto da posição do transmissor e incerteza de orientação na precisão
do sistema VLP com base na intensidade do sinal recebido (RSS) foi também investigado. Os resultados da simulação mostram que as incertezas do
transmissor têm um impacto severo no erro de posicionamento, que pode ser
atenuado com o uso de mais transmissores. Para incertezas de
posicionamento dos transmissores menores que 5 cm, os erros médios de
posicionamento são 23.3, 15.1 e 13.2 cm para conjuntos de 4, 9 e 16
transmissores, respetivamente. Enquanto que, para a incerteza de orientação
de um transmissor menor de 5°, os erros médios de posicionamento são 31.9,
20.6 e 17 cm para conjuntos de 4, 9 e 16 transmissores, respetivamente.
O trabalho da tese abordou a investigação dos aspetos de projeto de um
sistema VLP indoor no qual uma rede neuronal artificial (ANN) é utilizada para
estimativa de posicionamento considerando um canal multipercurso. O estudo
considerou a influência do ruído como indicador de desempenho para a
comparação entre diferentes abordagens de projeto. Três algoritmos de treino
de ANNs diferentes foram considerados, a saber, Levenberg-Marquardt,
regularização Bayesiana e algoritmos de gradiente conjugado escalonado,
para minimizar o erro de posicionamento no sistema VLP. O projeto da ANN foi
otimizado com base no número de neurónios nas camadas ocultas, no número
de épocas de treino e no tamanho do conjunto de treino. Mostrou-se que, a
ANN com regularização Bayesiana superou a técnica RSS tradicional usando
a estimação não linear dos mínimos quadrados para todos os valores da
relação sinal-ruído.
Foi proposto um novo sistema VLP indoor baseado em máquinas de vetores
de suporte (SVM) e regressão polinomial considerando dois ambientes
interiores diferentes: uma sala vazia e uma sala mobiliada. Os resultados
mostraram que, numa sala vazia, a melhoria da precisão de posicionamento
para o erro de posicionamento de 2.5 cm são 36.1, 58.3 e 72.2% para três
cenários diferentes de acordo com a distribuição das regiões na sala. Para a
sala mobiliada, uma melhoria de precisão relativa de posicionamento de 214,
170 e 100% é observada para erro de posicionamento de 0.1, 0.2 e 0.3 m,
respetivamente.
Finalmente, foi proposto um sistema VLP indoor baseado em redes neurais
convolucionais (CNN). O sistema foi demonstrado experimentalmente usando
luminárias LED como transmissores e uma camara com obturador rotativo
como recetor. O algoritmo de detecção usou um detector de disparo único
(SSD) baseado numa CNN pré configurada (ou seja, MobileNet ou ResNet)
para classificação. O sistema foi validado usando uma configuração de teste
de tamanho real contendo oito luminárias LED. Os resultados obtidos
mostraram que o erro de posicionamento quadrático médio alcançado é de
4.67 e 5.27 cm com os modelos SSD MobileNet e SSD ResNet,
respetivamente. Os resultados da validação mostram que o sistema pode
processar 67 imagens por segundo, permitindo o posicionamento em tempo
real.Programa Doutoral em Engenharia Eletrotécnic
Towards High-Frequency Tracking and Fast Edge-Aware Optimization
This dissertation advances the state of the art for AR/VR tracking systems by
increasing the tracking frequency by orders of magnitude and proposes an
efficient algorithm for the problem of edge-aware optimization.
AR/VR is a natural way of interacting with computers, where the physical and
digital worlds coexist. We are on the cusp of a radical change in how humans
perform and interact with computing. Humans are sensitive to small
misalignments between the real and the virtual world, and tracking at
kilo-Hertz frequencies becomes essential. Current vision-based systems fall
short, as their tracking frequency is implicitly limited by the frame-rate of
the camera. This thesis presents a prototype system which can track at orders
of magnitude higher than the state-of-the-art methods using multiple commodity
cameras. The proposed system exploits characteristics of the camera
traditionally considered as flaws, namely rolling shutter and radial
distortion. The experimental evaluation shows the effectiveness of the method
for various degrees of motion.
Furthermore, edge-aware optimization is an indispensable tool in the computer
vision arsenal for accurate filtering of depth-data and image-based rendering,
which is increasingly being used for content creation and geometry processing
for AR/VR. As applications increasingly demand higher resolution and speed,
there exists a need to develop methods that scale accordingly. This
dissertation proposes such an edge-aware optimization framework which is
efficient, accurate, and algorithmically scales well, all of which are much
desirable traits not found jointly in the state of the art. The experiments
show the effectiveness of the framework in a multitude of computer vision tasks
such as computational photography and stereo.Comment: PhD thesi
Dynamic spatial segmentation strategy based magnetic field indoor positioning system
In this day and age, it is imperative for anyone who relies on a mobile device to
track and navigate themselves using the Global Positioning System (GPS). Such
satellite-based positioning works as intended when in the outdoors, or when the
device is able to have unobstructed communication with GPS satellites.
Nevertheless, at the same time, GPS signal fades away in indoor environments due
to the effects of multi-path components and obstructed line-of-sight to the
satellite. Therefore, numerous indoor localisation applications have emerged in
the market, geared towards finding a practical solution to satisfy the need for
accuracy and efficiency.
The case of Indoor Positioning System (IPS) is promoted by recent smart devices,
which have evolved into a multimedia device with various sensors and optimised
connectivity. By sensing the device’s surroundings and inferring its context,
current IPS technology has proven its ability to provide stable and reliable indoor
localisation information. However, such a system is usually dependent on a high-density of infrastructure that requires expensive installations (e.g. Wi-Fi-based
IPS). To make a trade-off between accuracy and cost, considerable attention from
many researchers has been paid to the range of infrastructure-free technologies,
particularly exploiting the earth’s magnetic field (EMF).
EMF is a promising signal type that features ubiquitous availability, location
specificity and long-term stability. When considering the practicality of this
typical signal in IPS, such a system only consists of mobile device and the EMF
signal. To fully comprehend the conventional EMF-based IPS reported in the
literature, a preliminary experimental study on indoor EMF characteristics was
carried out at the beginning of this research. The results revealed that the positioning performance decreased when the presence of magnetic disturbance
sources was lowered to a minimum. In response to this finding, a new concept of
spatial segmentation is devised in this research based on magnetic anomaly (MA).
Therefore, this study focuses on developing innovative techniques based on spatial
segmentation strategy and machine learning algorithms for effective indoor
localisation using EMF.
In this thesis, four closely correlated components in the proposed system are
included: (i) Kriging interpolation-based fingerprinting map; (ii) magnetic
intensity-based spatial segmentation; (iii) weighted Naïve Bayes classification
(WNBC); (iv) fused features-based k-Nearest-Neighbours (kNN) algorithm.
Kriging interpolation-based fingerprinting map reconstructs the original observed
EMF positioning database in the calibration phase by interpolating predicted
points. The magnetic intensity-based spatial segmentation component then
investigates the variation tendency of ambient EMF signals in the new database to
analyse the distribution of magnetic disturbance sources, and accordingly,
segmenting the test site. Then, WNBC blends the exclusive characteristics of
indoor EMF into original Naïve Bayes Classification (NBC) to enable a more
accurate and efficient segmentation approach. It is well known that the best IPS
implementation often exerts the use of multiple positing sources in order to
maximise accuracy. The fused features-based kNN component used in the
positioning phase finally learns the various parameters collected in the calibration
phase, continuously improving the positioning accuracy of the system.
The proposed system was evaluated on multiple indoor sites with diverse layouts.
The results show that it outperforms state-of-the-art approaches and demonstrate
an average accuracy between 1-2 meters achieved in typical sites by the best
methods proposed in this thesis across most of the experimental environments. It
can be believed that such an accurate approach will enable the future of
infrastructure–free IPS technologies
Indoor Mapping and Reconstruction with Mobile Augmented Reality Sensor Systems
Augmented Reality (AR) ermöglicht es, virtuelle, dreidimensionale Inhalte direkt
innerhalb der realen Umgebung darzustellen. Anstatt jedoch beliebige virtuelle
Objekte an einem willkürlichen Ort anzuzeigen, kann AR Technologie auch genutzt
werden, um Geodaten in situ an jenem Ort darzustellen, auf den sich die Daten
beziehen. Damit eröffnet AR die Möglichkeit, die reale Welt durch virtuelle, ortbezogene
Informationen anzureichern. Im Rahmen der vorliegenen Arbeit wird diese
Spielart von AR als "Fused Reality" definiert und eingehend diskutiert.
Der praktische Mehrwert, den dieses Konzept der Fused Reality bietet, lässt sich
gut am Beispiel seiner Anwendung im Zusammenhang mit digitalen Gebäudemodellen
demonstrieren, wo sich gebäudespezifische Informationen - beispielsweise der
Verlauf von Leitungen und Kabeln innerhalb der Wände - lagegerecht am realen
Objekt darstellen lassen. Um das skizzierte Konzept einer Indoor Fused Reality
Anwendung realisieren zu können, müssen einige grundlegende Bedingungen erfüllt
sein. So kann ein bestimmtes Gebäude nur dann mit ortsbezogenen Informationen
augmentiert werden, wenn von diesem Gebäude ein digitales Modell verfügbar ist.
Zwar werden größere Bauprojekt heutzutage oft unter Zuhilfename von Building
Information Modelling (BIM) geplant und durchgeführt, sodass ein digitales Modell
direkt zusammen mit dem realen Gebäude ensteht, jedoch sind im Falle älterer
Bestandsgebäude digitale Modelle meist nicht verfügbar. Ein digitales Modell eines
bestehenden Gebäudes manuell zu erstellen, ist zwar möglich, jedoch mit großem
Aufwand verbunden. Ist ein passendes Gebäudemodell vorhanden, muss ein AR
Gerät außerdem in der Lage sein, die eigene Position und Orientierung im Gebäude
relativ zu diesem Modell bestimmen zu können, um Augmentierungen lagegerecht
anzeigen zu können.
Im Rahmen dieser Arbeit werden diverse Aspekte der angesprochenen Problematik
untersucht und diskutiert. Dabei werden zunächst verschiedene Möglichkeiten
diskutiert, Indoor-Gebäudegeometrie mittels Sensorsystemen zu erfassen. Anschließend
wird eine Untersuchung präsentiert, inwiefern moderne AR Geräte, die
in der Regel ebenfalls über eine Vielzahl an Sensoren verfügen, ebenfalls geeignet
sind, als Indoor-Mapping-Systeme eingesetzt zu werden. Die resultierenden Indoor
Mapping Datensätze können daraufhin genutzt werden, um automatisiert
Gebäudemodelle zu rekonstruieren. Zu diesem Zweck wird ein automatisiertes,
voxel-basiertes Indoor-Rekonstruktionsverfahren vorgestellt. Dieses wird außerdem
auf der Grundlage vierer zu diesem Zweck erfasster Datensätze mit zugehörigen
Referenzdaten quantitativ evaluiert. Desweiteren werden verschiedene
Möglichkeiten diskutiert, mobile AR Geräte innerhalb eines Gebäudes und des zugehörigen
Gebäudemodells zu lokalisieren. In diesem Kontext wird außerdem auch
die Evaluierung einer Marker-basierten Indoor-Lokalisierungsmethode präsentiert.
Abschließend wird zudem ein neuer Ansatz, Indoor-Mapping Datensätze an den
Achsen des Koordinatensystems auszurichten, vorgestellt
Structureless Camera Motion Estimation of Unordered Omnidirectional Images
This work aims at providing a novel camera motion estimation pipeline from large collections of unordered omnidirectional images. In oder to keep the pipeline as general and flexible as possible, cameras are modelled as unit spheres, allowing to incorporate any central camera type. For each camera an unprojection lookup is generated from intrinsics, which is called P2S-map (Pixel-to-Sphere-map), mapping pixels to their corresponding positions on the unit sphere. Consequently the camera geometry becomes independent of the underlying projection model. The pipeline also generates P2S-maps from world map projections with less distortion effects as they are known from cartography. Using P2S-maps from camera calibration and world map projection allows to convert omnidirectional camera images to an appropriate world map projection in oder to apply standard feature extraction and matching algorithms for data association. The proposed estimation pipeline combines the flexibility of SfM (Structure from Motion) - which handles unordered image collections - with the efficiency of PGO (Pose Graph Optimization), which is used as back-end in graph-based Visual SLAM (Simultaneous Localization and Mapping) approaches to optimize camera poses from large image sequences. SfM uses BA (Bundle Adjustment) to jointly optimize camera poses (motion) and 3d feature locations (structure), which becomes computationally expensive for large-scale scenarios. On the contrary PGO solves for camera poses (motion) from measured transformations between cameras, maintaining optimization managable. The proposed estimation algorithm combines both worlds. It obtains up-to-scale transformations between image pairs using two-view constraints, which are jointly scaled using trifocal constraints. A pose graph is generated from scaled two-view transformations and solved by PGO to obtain camera motion efficiently even for large image collections. Obtained results can be used as input data to provide initial pose estimates for further 3d reconstruction purposes e.g. to build a sparse structure from feature correspondences in an SfM or SLAM framework with further refinement via BA.
The pipeline also incorporates fixed extrinsic constraints from multi-camera setups as well as depth information provided by RGBD sensors. The entire camera motion estimation pipeline does not need to generate a sparse 3d structure of the captured environment and thus is called SCME (Structureless Camera Motion Estimation).:1 Introduction
1.1 Motivation
1.1.1 Increasing Interest of Image-Based 3D Reconstruction
1.1.2 Underground Environments as Challenging Scenario
1.1.3 Improved Mobile Camera Systems for Full Omnidirectional Imaging
1.2 Issues
1.2.1 Directional versus Omnidirectional Image Acquisition
1.2.2 Structure from Motion versus Visual Simultaneous Localization and Mapping
1.3 Contribution
1.4 Structure of this Work
2 Related Work
2.1 Visual Simultaneous Localization and Mapping
2.1.1 Visual Odometry
2.1.2 Pose Graph Optimization
2.2 Structure from Motion
2.2.1 Bundle Adjustment
2.2.2 Structureless Bundle Adjustment
2.3 Corresponding Issues
2.4 Proposed Reconstruction Pipeline
3 Cameras and Pixel-to-Sphere Mappings with P2S-Maps
3.1 Types
3.2 Models
3.2.1 Unified Camera Model
3.2.2 Polynomal Camera Model
3.2.3 Spherical Camera Model
3.3 P2S-Maps - Mapping onto Unit Sphere via Lookup Table
3.3.1 Lookup Table as Color Image
3.3.2 Lookup Interpolation
3.3.3 Depth Data Conversion
4 Calibration
4.1 Overview of Proposed Calibration Pipeline
4.2 Target Detection
4.3 Intrinsic Calibration
4.3.1 Selected Examples
4.4 Extrinsic Calibration
4.4.1 3D-2D Pose Estimation
4.4.2 2D-2D Pose Estimation
4.4.3 Pose Optimization
4.4.4 Uncertainty Estimation
4.4.5 PoseGraph Representation
4.4.6 Bundle Adjustment
4.4.7 Selected Examples
5 Full Omnidirectional Image Projections
5.1 Panoramic Image Stitching
5.2 World Map Projections
5.3 World Map Projection Generator for P2S-Maps
5.4 Conversion between Projections based on P2S-Maps
5.4.1 Proposed Workflow
5.4.2 Data Storage Format
5.4.3 Real World Example
6 Relations between Two Camera Spheres
6.1 Forward and Backward Projection
6.2 Triangulation
6.2.1 Linear Least Squares Method
6.2.2 Alternative Midpoint Method
6.3 Epipolar Geometry
6.4 Transformation Recovery from Essential Matrix
6.4.1 Cheirality
6.4.2 Standard Procedure
6.4.3 Simplified Procedure
6.4.4 Improved Procedure
6.5 Two-View Estimation
6.5.1 Evaluation Strategy
6.5.2 Error Metric
6.5.3 Evaluation of Estimation Algorithms
6.5.4 Concluding Remarks
6.6 Two-View Optimization
6.6.1 Epipolar-Based Error Distances
6.6.2 Projection-Based Error Distances
6.6.3 Comparison between Error Distances
6.7 Two-View Translation Scaling
6.7.1 Linear Least Squares Estimation
6.7.2 Non-Linear Least Squares Optimization
6.7.3 Comparison between Initial and Optimized Scaling Factor
6.8 Homography to Identify Degeneracies
6.8.1 Homography for Spherical Cameras
6.8.2 Homography Estimation
6.8.3 Homography Optimization
6.8.4 Homography and Pure Rotation
6.8.5 Homography in Epipolar Geometry
7 Relations between Three Camera Spheres
7.1 Three View Geometry
7.2 Crossing Epipolar Planes Geometry
7.3 Trifocal Geometry
7.4 Relation between Trifocal, Three-View and Crossing Epipolar Planes
7.5 Translation Ratio between Up-To-Scale Two-View Transformations
7.5.1 Structureless Determination Approaches
7.5.2 Structure-Based Determination Approaches
7.5.3 Comparison between Proposed Approaches
8 Pose Graphs
8.1 Optimization Principle
8.2 Solvers
8.2.1 Additional Graph Solvers
8.2.2 False Loop Closure Detection
8.3 Pose Graph Generation
8.3.1 Generation of Synthetic Pose Graph Data
8.3.2 Optimization of Synthetic Pose Graph Data
9 Structureless Camera Motion Estimation
9.1 SCME Pipeline
9.2 Determination of Two-View Translation Scale Factors
9.3 Integration of Depth Data
9.4 Integration of Extrinsic Camera Constraints
10 Camera Motion Estimation Results
10.1 Directional Camera Images
10.2 Omnidirectional Camera Images
11 Conclusion
11.1 Summary
11.2 Outlook and Future Work
Appendices
A.1 Additional Extrinsic Calibration Results
A.2 Linear Least Squares Scaling
A.3 Proof Rank Deficiency
A.4 Alternative Derivation Midpoint Method
A.5 Simplification of Depth Calculation
A.6 Relation between Epipolar and Circumferential Constraint
A.7 Covariance Estimation
A.8 Uncertainty Estimation from Epipolar Geometry
A.9 Two-View Scaling Factor Estimation: Uncertainty Estimation
A.10 Two-View Scaling Factor Optimization: Uncertainty Estimation
A.11 Depth from Adjoining Two-View Geometries
A.12 Alternative Three-View Derivation
A.12.1 Second Derivation Approach
A.12.2 Third Derivation Approach
A.13 Relation between Trifocal Geometry and Alternative Midpoint Method
A.14 Additional Pose Graph Generation Examples
A.15 Pose Graph Solver Settings
A.16 Additional Pose Graph Optimization Examples
Bibliograph
Systems with Massive Number of Antennas: Distributed Approaches
As 5G is entering maturity, the research interest has shifted towards 6G, and specially the new use cases that the future telecommunication infrastructure needs to support. These new use cases encompass much higher requirements, specifically: higher communication data-rates, larger number of users, higher accuracy in localization, possibility to wirelessly charge devices, among others.The radio access network (RAN) has already gone through an evolution on the path towards 5G. One of the main changes was a large increment of the number of antennas in the base-station. Some of them may even reach 100 elements, in what is commonly referred as Massive MIMO. New proposals for 6G RAN point in the direction of continuing this path of increasing the number of antennas, and locate them throughout a certain area of service. Different technologies have been proposed in this direction, such as: cell-free Massive MIMO, distributed MIMO, and large intelligent surface (LIS). In this thesis we focus on LIS, whose conducted theoretical studies promise the fulfillment of the aforementioned requirements.While the theoretical capabilities of LIS have been conveniently analyzed, little has been done in terms of implementing this type of systems. When the number of antennas grow to hundreds or thousands, there are numerous challenges that need to be solved for a successful implementation. The most critical challenges are the interconnection data-rate and the computational complexity.In the present thesis we introduce the implementation challenges, and show that centralized processing architectures are no longer adequate for this type of systems. We also present different distributed processing architectures and show the benefits of this type of schemes. This work aims at giving a system-design guideline that helps the system designer to make the right decisions when designing these type of systems. For that, we provide algorithms, performance analysis and comparisons, including first order evaluation of the interconnection data-rate, processing latency, memory and energy consumption. These numbers are based on models and available data in the literature. Exact values depend on the selected technology, and will be accurately determined after building and testing these type of systems.The thesis concentrates mostly on the topic of communication, with additional exploration of other areas, such as localization. In case of localization, we benefit from the high spatial resolution of a very-large array that provides very rich channel state information (CSI). A CSI-based fingerprinting via neural network technique is selected for this case with promising results. As the communication and localization services are based on the acquisition of CSI, we foresee a common system architecture capable of supporting both cases. Further work in this direction is recommended, with the possibility of including other applications such as sensing.The obtained results indicate that the implementation of these very-large array systems is feasible, but the challenges are numerous. The proposed solutions provide encouraging results that need to be verified with hardware implementations and real measurements