44 research outputs found

    MEMS Accelerometers

    Get PDF
    Micro-electro-mechanical system (MEMS) devices are widely used for inertia, pressure, and ultrasound sensing applications. Research on integrated MEMS technology has undergone extensive development driven by the requirements of a compact footprint, low cost, and increased functionality. Accelerometers are among the most widely used sensors implemented in MEMS technology. MEMS accelerometers are showing a growing presence in almost all industries ranging from automotive to medical. A traditional MEMS accelerometer employs a proof mass suspended to springs, which displaces in response to an external acceleration. A single proof mass can be used for one- or multi-axis sensing. A variety of transduction mechanisms have been used to detect the displacement. They include capacitive, piezoelectric, thermal, tunneling, and optical mechanisms. Capacitive accelerometers are widely used due to their DC measurement interface, thermal stability, reliability, and low cost. However, they are sensitive to electromagnetic field interferences and have poor performance for high-end applications (e.g., precise attitude control for the satellite). Over the past three decades, steady progress has been made in the area of optical accelerometers for high-performance and high-sensitivity applications but several challenges are still to be tackled by researchers and engineers to fully realize opto-mechanical accelerometers, such as chip-scale integration, scaling, low bandwidth, etc

    Sistemas de posicionamento baseados em comunicação por luz para ambientes interiores

    Get PDF
    The demand for highly precise indoor positioning systems (IPSs) is growing rapidly due to its potential in the increasingly popular techniques of the Internet of Things, smart mobile devices, and artificial intelligence. IPS becomes a promising research domain that is getting wide attention due to its benefits in several working scenarios, such as, industries, indoor public locations, and autonomous navigation. Moreover, IPS has a prominent contribution in day-to-day activities in organizations such as health care centers, airports, shopping malls, manufacturing, underground locations, etc., for safe operating environments. In indoor environments, both radio frequency (RF) and optical wireless communication (OWC) based technologies could be adopted for localization. Although the RF-based global positioning system, such as, Global positioning system offers higher penetration rates with reduced accuracy (i.e., in the range of a few meters), it does not work well in indoor environments (and not at all in certain cases such as tunnels, mines, etc.) due to the very weak signal and no direct access to the satellites. On the other hand, the light-based system known as a visible light positioning (VLP) system, as part of the OWC systems, uses the pre-existing light-emitting diodes (LEDs)-based lighting infrastructure, could be used at low cost and high accuracy compared with the RF-based systems. VLP is an emerging technology promising high accuracy, high security, low deployment cost, shorter time response, and low relative complexity when compared with RFbased positioning. However, in indoor VLP systems, there are some concerns such as, multipath reflection, transmitter tilting, transmitter’s position, and orientation uncertainty, human shadowing/blocking, and noise causing the increase in the positioning error, thereby reducing the positioning accuracy of the system. Therefore, it is imperative to capture the characteristics of different VLP channel and properly model them for the dual purpose of illumination and localization. In this thesis, firstly, the impact of transmitter tilting angles and multipath reflections are studied and for the first time, it is demonstrated that tilting the transmitter can be beneficial in VLP systems considering both line of sight (LOS) and non-line of sight transmission paths. With the transmitters oriented towards the center of the receiving plane, the received power level is maximized due to the LOS components. It is also shown that the proposed scheme offers a significant accuracy improvement of up to ~66% compared with a typical non-tilted transmitter VLP. The effect of tilting the transmitter on the lighting uniformity is also investigated and results proved that the uniformity achieved complies with the European Standard EN 12464-1. After that, the impact of transmitter position and orientation uncertainty on the accuracy of the VLP system based on the received signal strength (RSS) is investigated. Simulation results show that the transmitter uncertainties have a severe impact on the positioning error, which can be leveraged through the usage of more transmitters. Concerning a smaller transmitter’s position epochs, and the size of the training set. It is shown that, the ANN with Bayesian regularization outperforms the traditional RSS technique using the non-linear least square estimation for all values of signal to noise ratio. Furthermore, a novel indoor VLP system is proposed based on support vector machines and polynomial regression considering two different multipath environments of an empty room and a furnished room. The results show that, in an empty room, the positioning accuracy improvement for the positioning error of 2.5 cm are 36.1, 58.3, and 72.2 % for three different scenarios according to the regions’ distribution in the room. For the furnished room, a positioning relative accuracy improvement of 214, 170, and 100 % is observed for positioning error of 0.1, 0.2, and 0.3 m, respectively. Ultimately, an indoor VLP system based on convolutional neural networks (CNN) is proposed and demonstrated experimentally in which LEDs are used as transmitters and a rolling shutter camera is used as receiver. A detection algorithm named single shot detector (SSD) is used which relies on CNN (i.e., MobileNet or ResNet) for classification as well as position estimation of each LED in the image. The system is validated using a real-world size test setup containing eight LED luminaries. The obtained results show that the maximum average root mean square positioning error achieved is 4.67 and 5.27 cm with SSD MobileNet and SSD ResNet models, respectively. The validation results show that the system can process 67 images per second, allowing real-time positioning.A procura por sistemas de posicionamento interior (IPSs) de alta precisão tem crescido rapidamente devido ao seu interesse nas técnicas cada vez mais populares da Internet das Coisas, dispositivos móveis inteligentes e inteligência artificial. O IPS tornou-se um domínio de pesquisa promissor que tem atraído grande atenção devido aos seus benefícios em vários cenários de trabalho, como indústrias, locais públicos e navegação autónoma. Além disso, o IPS tem uma contribuição destacada no dia a dia de organizações, como, centros de saúde, aeroportos, supermercados, fábricas, locais subterrâneos, etc. As tecnologias baseadas em radiofrequência (RF) e comunicação óptica sem fio (OWC) podem ser adotadas para localização em ambientes interiores. Embora o sistema de posicionamento global (GPS) baseado em RF ofereça taxas de penetração mais altas com precisão reduzida (ou seja, na faixa de alguns metros), não funciona bem em ambientes interiores (e não funciona bem em certos casos como túneis, minas, etc.) devido ao sinal muito fraco e falta de acesso direto aos satélites. Por outro lado, o sistema baseado em luz conhecido como sistema de posicionamento de luz visível (VLP), como parte dos sistemas OWC, usa a infraestrutura de iluminação baseada em díodos emissores de luz (LEDs) pré-existentes, é um sistemas de baixo custo e alta precisão quando comprado com os sistemas baseados em RF. O VLP é uma tecnologia emergente que promete alta precisão, alta segurança, baixo custo de implantação, menor tempo de resposta e baixa complexidade relativa quando comparado ao posicionamento baseado em RF. No entanto, os sistemas VLP interiores, exibem algumas limitações, como, a reflexão multicaminho, inclinação do transmissor, posição do transmissor e incerteza de orientação, sombra/bloqueio humano e ruído, que têm como consequência o aumento do erro de posicionamento, e consequente redução da precisão do sistema. Portanto, é imperativo estudar as características dos diferentes canais VLP e modelá-los adequadamente para o duplo propósito de iluminação e localização. Esta tesa aborda, primeiramente, o impacto dos ângulos de inclinação do transmissor e reflexões multipercurso no desempenho do sistema de posicionamento. Demonstra-se que a inclinação do transmissor pode ser benéfica em sistemas VLP considerando tanto a linha de vista (LOS) como as reflexões. Com os transmissores orientados para o centro do plano recetor, o nível de potência recebido é maximizado devido aos componentes LOS. Também é mostrado que o esquema proposto oferece uma melhoria significativa de precisão de até ~66% em comparação com um sistema VLP de transmissor não inclinado típico. O efeito da inclinação do transmissor na uniformidade da iluminação também é investigado e os resultados comprovam que a uniformidade alcançada está de acordo com a Norma Europeia EN 12464-1. O impacto da posição do transmissor e incerteza de orientação na precisão do sistema VLP com base na intensidade do sinal recebido (RSS) foi também investigado. Os resultados da simulação mostram que as incertezas do transmissor têm um impacto severo no erro de posicionamento, que pode ser atenuado com o uso de mais transmissores. Para incertezas de posicionamento dos transmissores menores que 5 cm, os erros médios de posicionamento são 23.3, 15.1 e 13.2 cm para conjuntos de 4, 9 e 16 transmissores, respetivamente. Enquanto que, para a incerteza de orientação de um transmissor menor de 5°, os erros médios de posicionamento são 31.9, 20.6 e 17 cm para conjuntos de 4, 9 e 16 transmissores, respetivamente. O trabalho da tese abordou a investigação dos aspetos de projeto de um sistema VLP indoor no qual uma rede neuronal artificial (ANN) é utilizada para estimativa de posicionamento considerando um canal multipercurso. O estudo considerou a influência do ruído como indicador de desempenho para a comparação entre diferentes abordagens de projeto. Três algoritmos de treino de ANNs diferentes foram considerados, a saber, Levenberg-Marquardt, regularização Bayesiana e algoritmos de gradiente conjugado escalonado, para minimizar o erro de posicionamento no sistema VLP. O projeto da ANN foi otimizado com base no número de neurónios nas camadas ocultas, no número de épocas de treino e no tamanho do conjunto de treino. Mostrou-se que, a ANN com regularização Bayesiana superou a técnica RSS tradicional usando a estimação não linear dos mínimos quadrados para todos os valores da relação sinal-ruído. Foi proposto um novo sistema VLP indoor baseado em máquinas de vetores de suporte (SVM) e regressão polinomial considerando dois ambientes interiores diferentes: uma sala vazia e uma sala mobiliada. Os resultados mostraram que, numa sala vazia, a melhoria da precisão de posicionamento para o erro de posicionamento de 2.5 cm são 36.1, 58.3 e 72.2% para três cenários diferentes de acordo com a distribuição das regiões na sala. Para a sala mobiliada, uma melhoria de precisão relativa de posicionamento de 214, 170 e 100% é observada para erro de posicionamento de 0.1, 0.2 e 0.3 m, respetivamente. Finalmente, foi proposto um sistema VLP indoor baseado em redes neurais convolucionais (CNN). O sistema foi demonstrado experimentalmente usando luminárias LED como transmissores e uma camara com obturador rotativo como recetor. O algoritmo de detecção usou um detector de disparo único (SSD) baseado numa CNN pré configurada (ou seja, MobileNet ou ResNet) para classificação. O sistema foi validado usando uma configuração de teste de tamanho real contendo oito luminárias LED. Os resultados obtidos mostraram que o erro de posicionamento quadrático médio alcançado é de 4.67 e 5.27 cm com os modelos SSD MobileNet e SSD ResNet, respetivamente. Os resultados da validação mostram que o sistema pode processar 67 imagens por segundo, permitindo o posicionamento em tempo real.Programa Doutoral em Engenharia Eletrotécnic

    Towards High-Frequency Tracking and Fast Edge-Aware Optimization

    Full text link
    This dissertation advances the state of the art for AR/VR tracking systems by increasing the tracking frequency by orders of magnitude and proposes an efficient algorithm for the problem of edge-aware optimization. AR/VR is a natural way of interacting with computers, where the physical and digital worlds coexist. We are on the cusp of a radical change in how humans perform and interact with computing. Humans are sensitive to small misalignments between the real and the virtual world, and tracking at kilo-Hertz frequencies becomes essential. Current vision-based systems fall short, as their tracking frequency is implicitly limited by the frame-rate of the camera. This thesis presents a prototype system which can track at orders of magnitude higher than the state-of-the-art methods using multiple commodity cameras. The proposed system exploits characteristics of the camera traditionally considered as flaws, namely rolling shutter and radial distortion. The experimental evaluation shows the effectiveness of the method for various degrees of motion. Furthermore, edge-aware optimization is an indispensable tool in the computer vision arsenal for accurate filtering of depth-data and image-based rendering, which is increasingly being used for content creation and geometry processing for AR/VR. As applications increasingly demand higher resolution and speed, there exists a need to develop methods that scale accordingly. This dissertation proposes such an edge-aware optimization framework which is efficient, accurate, and algorithmically scales well, all of which are much desirable traits not found jointly in the state of the art. The experiments show the effectiveness of the framework in a multitude of computer vision tasks such as computational photography and stereo.Comment: PhD thesi

    Intelligent Sensing and Learning for Advanced MIMO Communication Systems

    Get PDF

    Dynamic spatial segmentation strategy based magnetic field indoor positioning system

    Get PDF
    In this day and age, it is imperative for anyone who relies on a mobile device to track and navigate themselves using the Global Positioning System (GPS). Such satellite-based positioning works as intended when in the outdoors, or when the device is able to have unobstructed communication with GPS satellites. Nevertheless, at the same time, GPS signal fades away in indoor environments due to the effects of multi-path components and obstructed line-of-sight to the satellite. Therefore, numerous indoor localisation applications have emerged in the market, geared towards finding a practical solution to satisfy the need for accuracy and efficiency. The case of Indoor Positioning System (IPS) is promoted by recent smart devices, which have evolved into a multimedia device with various sensors and optimised connectivity. By sensing the device’s surroundings and inferring its context, current IPS technology has proven its ability to provide stable and reliable indoor localisation information. However, such a system is usually dependent on a high-density of infrastructure that requires expensive installations (e.g. Wi-Fi-based IPS). To make a trade-off between accuracy and cost, considerable attention from many researchers has been paid to the range of infrastructure-free technologies, particularly exploiting the earth’s magnetic field (EMF). EMF is a promising signal type that features ubiquitous availability, location specificity and long-term stability. When considering the practicality of this typical signal in IPS, such a system only consists of mobile device and the EMF signal. To fully comprehend the conventional EMF-based IPS reported in the literature, a preliminary experimental study on indoor EMF characteristics was carried out at the beginning of this research. The results revealed that the positioning performance decreased when the presence of magnetic disturbance sources was lowered to a minimum. In response to this finding, a new concept of spatial segmentation is devised in this research based on magnetic anomaly (MA). Therefore, this study focuses on developing innovative techniques based on spatial segmentation strategy and machine learning algorithms for effective indoor localisation using EMF. In this thesis, four closely correlated components in the proposed system are included: (i) Kriging interpolation-based fingerprinting map; (ii) magnetic intensity-based spatial segmentation; (iii) weighted Naïve Bayes classification (WNBC); (iv) fused features-based k-Nearest-Neighbours (kNN) algorithm. Kriging interpolation-based fingerprinting map reconstructs the original observed EMF positioning database in the calibration phase by interpolating predicted points. The magnetic intensity-based spatial segmentation component then investigates the variation tendency of ambient EMF signals in the new database to analyse the distribution of magnetic disturbance sources, and accordingly, segmenting the test site. Then, WNBC blends the exclusive characteristics of indoor EMF into original Naïve Bayes Classification (NBC) to enable a more accurate and efficient segmentation approach. It is well known that the best IPS implementation often exerts the use of multiple positing sources in order to maximise accuracy. The fused features-based kNN component used in the positioning phase finally learns the various parameters collected in the calibration phase, continuously improving the positioning accuracy of the system. The proposed system was evaluated on multiple indoor sites with diverse layouts. The results show that it outperforms state-of-the-art approaches and demonstrate an average accuracy between 1-2 meters achieved in typical sites by the best methods proposed in this thesis across most of the experimental environments. It can be believed that such an accurate approach will enable the future of infrastructure–free IPS technologies

    Indoor Mapping and Reconstruction with Mobile Augmented Reality Sensor Systems

    Get PDF
    Augmented Reality (AR) ermöglicht es, virtuelle, dreidimensionale Inhalte direkt innerhalb der realen Umgebung darzustellen. Anstatt jedoch beliebige virtuelle Objekte an einem willkürlichen Ort anzuzeigen, kann AR Technologie auch genutzt werden, um Geodaten in situ an jenem Ort darzustellen, auf den sich die Daten beziehen. Damit eröffnet AR die Möglichkeit, die reale Welt durch virtuelle, ortbezogene Informationen anzureichern. Im Rahmen der vorliegenen Arbeit wird diese Spielart von AR als "Fused Reality" definiert und eingehend diskutiert. Der praktische Mehrwert, den dieses Konzept der Fused Reality bietet, lässt sich gut am Beispiel seiner Anwendung im Zusammenhang mit digitalen Gebäudemodellen demonstrieren, wo sich gebäudespezifische Informationen - beispielsweise der Verlauf von Leitungen und Kabeln innerhalb der Wände - lagegerecht am realen Objekt darstellen lassen. Um das skizzierte Konzept einer Indoor Fused Reality Anwendung realisieren zu können, müssen einige grundlegende Bedingungen erfüllt sein. So kann ein bestimmtes Gebäude nur dann mit ortsbezogenen Informationen augmentiert werden, wenn von diesem Gebäude ein digitales Modell verfügbar ist. Zwar werden größere Bauprojekt heutzutage oft unter Zuhilfename von Building Information Modelling (BIM) geplant und durchgeführt, sodass ein digitales Modell direkt zusammen mit dem realen Gebäude ensteht, jedoch sind im Falle älterer Bestandsgebäude digitale Modelle meist nicht verfügbar. Ein digitales Modell eines bestehenden Gebäudes manuell zu erstellen, ist zwar möglich, jedoch mit großem Aufwand verbunden. Ist ein passendes Gebäudemodell vorhanden, muss ein AR Gerät außerdem in der Lage sein, die eigene Position und Orientierung im Gebäude relativ zu diesem Modell bestimmen zu können, um Augmentierungen lagegerecht anzeigen zu können. Im Rahmen dieser Arbeit werden diverse Aspekte der angesprochenen Problematik untersucht und diskutiert. Dabei werden zunächst verschiedene Möglichkeiten diskutiert, Indoor-Gebäudegeometrie mittels Sensorsystemen zu erfassen. Anschließend wird eine Untersuchung präsentiert, inwiefern moderne AR Geräte, die in der Regel ebenfalls über eine Vielzahl an Sensoren verfügen, ebenfalls geeignet sind, als Indoor-Mapping-Systeme eingesetzt zu werden. Die resultierenden Indoor Mapping Datensätze können daraufhin genutzt werden, um automatisiert Gebäudemodelle zu rekonstruieren. Zu diesem Zweck wird ein automatisiertes, voxel-basiertes Indoor-Rekonstruktionsverfahren vorgestellt. Dieses wird außerdem auf der Grundlage vierer zu diesem Zweck erfasster Datensätze mit zugehörigen Referenzdaten quantitativ evaluiert. Desweiteren werden verschiedene Möglichkeiten diskutiert, mobile AR Geräte innerhalb eines Gebäudes und des zugehörigen Gebäudemodells zu lokalisieren. In diesem Kontext wird außerdem auch die Evaluierung einer Marker-basierten Indoor-Lokalisierungsmethode präsentiert. Abschließend wird zudem ein neuer Ansatz, Indoor-Mapping Datensätze an den Achsen des Koordinatensystems auszurichten, vorgestellt

    Structureless Camera Motion Estimation of Unordered Omnidirectional Images

    Get PDF
    This work aims at providing a novel camera motion estimation pipeline from large collections of unordered omnidirectional images. In oder to keep the pipeline as general and flexible as possible, cameras are modelled as unit spheres, allowing to incorporate any central camera type. For each camera an unprojection lookup is generated from intrinsics, which is called P2S-map (Pixel-to-Sphere-map), mapping pixels to their corresponding positions on the unit sphere. Consequently the camera geometry becomes independent of the underlying projection model. The pipeline also generates P2S-maps from world map projections with less distortion effects as they are known from cartography. Using P2S-maps from camera calibration and world map projection allows to convert omnidirectional camera images to an appropriate world map projection in oder to apply standard feature extraction and matching algorithms for data association. The proposed estimation pipeline combines the flexibility of SfM (Structure from Motion) - which handles unordered image collections - with the efficiency of PGO (Pose Graph Optimization), which is used as back-end in graph-based Visual SLAM (Simultaneous Localization and Mapping) approaches to optimize camera poses from large image sequences. SfM uses BA (Bundle Adjustment) to jointly optimize camera poses (motion) and 3d feature locations (structure), which becomes computationally expensive for large-scale scenarios. On the contrary PGO solves for camera poses (motion) from measured transformations between cameras, maintaining optimization managable. The proposed estimation algorithm combines both worlds. It obtains up-to-scale transformations between image pairs using two-view constraints, which are jointly scaled using trifocal constraints. A pose graph is generated from scaled two-view transformations and solved by PGO to obtain camera motion efficiently even for large image collections. Obtained results can be used as input data to provide initial pose estimates for further 3d reconstruction purposes e.g. to build a sparse structure from feature correspondences in an SfM or SLAM framework with further refinement via BA. The pipeline also incorporates fixed extrinsic constraints from multi-camera setups as well as depth information provided by RGBD sensors. The entire camera motion estimation pipeline does not need to generate a sparse 3d structure of the captured environment and thus is called SCME (Structureless Camera Motion Estimation).:1 Introduction 1.1 Motivation 1.1.1 Increasing Interest of Image-Based 3D Reconstruction 1.1.2 Underground Environments as Challenging Scenario 1.1.3 Improved Mobile Camera Systems for Full Omnidirectional Imaging 1.2 Issues 1.2.1 Directional versus Omnidirectional Image Acquisition 1.2.2 Structure from Motion versus Visual Simultaneous Localization and Mapping 1.3 Contribution 1.4 Structure of this Work 2 Related Work 2.1 Visual Simultaneous Localization and Mapping 2.1.1 Visual Odometry 2.1.2 Pose Graph Optimization 2.2 Structure from Motion 2.2.1 Bundle Adjustment 2.2.2 Structureless Bundle Adjustment 2.3 Corresponding Issues 2.4 Proposed Reconstruction Pipeline 3 Cameras and Pixel-to-Sphere Mappings with P2S-Maps 3.1 Types 3.2 Models 3.2.1 Unified Camera Model 3.2.2 Polynomal Camera Model 3.2.3 Spherical Camera Model 3.3 P2S-Maps - Mapping onto Unit Sphere via Lookup Table 3.3.1 Lookup Table as Color Image 3.3.2 Lookup Interpolation 3.3.3 Depth Data Conversion 4 Calibration 4.1 Overview of Proposed Calibration Pipeline 4.2 Target Detection 4.3 Intrinsic Calibration 4.3.1 Selected Examples 4.4 Extrinsic Calibration 4.4.1 3D-2D Pose Estimation 4.4.2 2D-2D Pose Estimation 4.4.3 Pose Optimization 4.4.4 Uncertainty Estimation 4.4.5 PoseGraph Representation 4.4.6 Bundle Adjustment 4.4.7 Selected Examples 5 Full Omnidirectional Image Projections 5.1 Panoramic Image Stitching 5.2 World Map Projections 5.3 World Map Projection Generator for P2S-Maps 5.4 Conversion between Projections based on P2S-Maps 5.4.1 Proposed Workflow 5.4.2 Data Storage Format 5.4.3 Real World Example 6 Relations between Two Camera Spheres 6.1 Forward and Backward Projection 6.2 Triangulation 6.2.1 Linear Least Squares Method 6.2.2 Alternative Midpoint Method 6.3 Epipolar Geometry 6.4 Transformation Recovery from Essential Matrix 6.4.1 Cheirality 6.4.2 Standard Procedure 6.4.3 Simplified Procedure 6.4.4 Improved Procedure 6.5 Two-View Estimation 6.5.1 Evaluation Strategy 6.5.2 Error Metric 6.5.3 Evaluation of Estimation Algorithms 6.5.4 Concluding Remarks 6.6 Two-View Optimization 6.6.1 Epipolar-Based Error Distances 6.6.2 Projection-Based Error Distances 6.6.3 Comparison between Error Distances 6.7 Two-View Translation Scaling 6.7.1 Linear Least Squares Estimation 6.7.2 Non-Linear Least Squares Optimization 6.7.3 Comparison between Initial and Optimized Scaling Factor 6.8 Homography to Identify Degeneracies 6.8.1 Homography for Spherical Cameras 6.8.2 Homography Estimation 6.8.3 Homography Optimization 6.8.4 Homography and Pure Rotation 6.8.5 Homography in Epipolar Geometry 7 Relations between Three Camera Spheres 7.1 Three View Geometry 7.2 Crossing Epipolar Planes Geometry 7.3 Trifocal Geometry 7.4 Relation between Trifocal, Three-View and Crossing Epipolar Planes 7.5 Translation Ratio between Up-To-Scale Two-View Transformations 7.5.1 Structureless Determination Approaches 7.5.2 Structure-Based Determination Approaches 7.5.3 Comparison between Proposed Approaches 8 Pose Graphs 8.1 Optimization Principle 8.2 Solvers 8.2.1 Additional Graph Solvers 8.2.2 False Loop Closure Detection 8.3 Pose Graph Generation 8.3.1 Generation of Synthetic Pose Graph Data 8.3.2 Optimization of Synthetic Pose Graph Data 9 Structureless Camera Motion Estimation 9.1 SCME Pipeline 9.2 Determination of Two-View Translation Scale Factors 9.3 Integration of Depth Data 9.4 Integration of Extrinsic Camera Constraints 10 Camera Motion Estimation Results 10.1 Directional Camera Images 10.2 Omnidirectional Camera Images 11 Conclusion 11.1 Summary 11.2 Outlook and Future Work Appendices A.1 Additional Extrinsic Calibration Results A.2 Linear Least Squares Scaling A.3 Proof Rank Deficiency A.4 Alternative Derivation Midpoint Method A.5 Simplification of Depth Calculation A.6 Relation between Epipolar and Circumferential Constraint A.7 Covariance Estimation A.8 Uncertainty Estimation from Epipolar Geometry A.9 Two-View Scaling Factor Estimation: Uncertainty Estimation A.10 Two-View Scaling Factor Optimization: Uncertainty Estimation A.11 Depth from Adjoining Two-View Geometries A.12 Alternative Three-View Derivation A.12.1 Second Derivation Approach A.12.2 Third Derivation Approach A.13 Relation between Trifocal Geometry and Alternative Midpoint Method A.14 Additional Pose Graph Generation Examples A.15 Pose Graph Solver Settings A.16 Additional Pose Graph Optimization Examples Bibliograph

    Systems with Massive Number of Antennas: Distributed Approaches

    Get PDF
    As 5G is entering maturity, the research interest has shifted towards 6G, and specially the new use cases that the future telecommunication infrastructure needs to support. These new use cases encompass much higher requirements, specifically: higher communication data-rates, larger number of users, higher accuracy in localization, possibility to wirelessly charge devices, among others.The radio access network (RAN) has already gone through an evolution on the path towards 5G. One of the main changes was a large increment of the number of antennas in the base-station. Some of them may even reach 100 elements, in what is commonly referred as Massive MIMO. New proposals for 6G RAN point in the direction of continuing this path of increasing the number of antennas, and locate them throughout a certain area of service. Different technologies have been proposed in this direction, such as: cell-free Massive MIMO, distributed MIMO, and large intelligent surface (LIS). In this thesis we focus on LIS, whose conducted theoretical studies promise the fulfillment of the aforementioned requirements.While the theoretical capabilities of LIS have been conveniently analyzed, little has been done in terms of implementing this type of systems. When the number of antennas grow to hundreds or thousands, there are numerous challenges that need to be solved for a successful implementation. The most critical challenges are the interconnection data-rate and the computational complexity.In the present thesis we introduce the implementation challenges, and show that centralized processing architectures are no longer adequate for this type of systems. We also present different distributed processing architectures and show the benefits of this type of schemes. This work aims at giving a system-design guideline that helps the system designer to make the right decisions when designing these type of systems. For that, we provide algorithms, performance analysis and comparisons, including first order evaluation of the interconnection data-rate, processing latency, memory and energy consumption. These numbers are based on models and available data in the literature. Exact values depend on the selected technology, and will be accurately determined after building and testing these type of systems.The thesis concentrates mostly on the topic of communication, with additional exploration of other areas, such as localization. In case of localization, we benefit from the high spatial resolution of a very-large array that provides very rich channel state information (CSI). A CSI-based fingerprinting via neural network technique is selected for this case with promising results. As the communication and localization services are based on the acquisition of CSI, we foresee a common system architecture capable of supporting both cases. Further work in this direction is recommended, with the possibility of including other applications such as sensing.The obtained results indicate that the implementation of these very-large array systems is feasible, but the challenges are numerous. The proposed solutions provide encouraging results that need to be verified with hardware implementations and real measurements
    corecore