2,248 research outputs found

    Efficient rendering for three-dimensional displays

    Get PDF
    This thesis explores more efficient methods for visualizing point data sets on three-dimensional (3D) displays. Point data sets are used in many scientific applications, e.g. cosmological simulations. Visualizing these data sets in {3D} is desirable because it can more readily reveal structure and unknown phenomena. However, cutting-edge scientific point data sets are very large and producing/rendering even a single image is expensive. Furthermore, current literature suggests that the ideal number of views for 3D (multiview) displays can be in the hundreds, which compounds the costs. The accepted notion that many views are required for {3D} displays is challenged by carrying out a novel human factor trials study. The results suggest that humans are actually surprisingly insensitive to the number of viewpoints with regard to their task performance, when occlusion in the scene is not a dominant factor. Existing stereoscopic rendering algorithms can have high set-up costs which limits their use and none are tuned for uncorrelated {3D} point rendering. This thesis shows that it is possible to improve rendering speeds for a low number of views by perspective reprojection. The novelty in the approach described lies in delaying the reprojection and generation of the viewpoints until the fragment stage of the pipeline and streamlining the rendering pipeline for points only. Theoretical analysis suggests a fragment reprojection scheme will render at least 2.8 times faster than na\"{i}vely re-rendering the scene from multiple viewpoints. Building upon the fragment reprojection technique, further rendering performance is shown to be possible (at the cost of some rendering accuracy) by restricting the amount of reprojection required according to the stereoscopic resolution of the display. A significant benefit is that the scene depth can be mapped arbitrarily to the perceived depth range of the display at no extra cost than a single region mapping approach. Using an average case-study (rendering from a 500k points for a 9-view High Definition 3D display), theoretical analysis suggests that this new approach is capable of twice the performance gains than simply reprojecting every single fragment, and quantitative measures show the algorithm to be 5 times faster than a naĂŻve rendering approach. Further detailed quantitative results, under varying scenarios, are provided and discussed

    Smart environment monitoring through micro unmanned aerial vehicles

    Get PDF
    In recent years, the improvements of small-scale Unmanned Aerial Vehicles (UAVs) in terms of flight time, automatic control, and remote transmission are promoting the development of a wide range of practical applications. In aerial video surveillance, the monitoring of broad areas still has many challenges due to the achievement of different tasks in real-time, including mosaicking, change detection, and object detection. In this thesis work, a small-scale UAV based vision system to maintain regular surveillance over target areas is proposed. The system works in two modes. The first mode allows to monitor an area of interest by performing several flights. During the first flight, it creates an incremental geo-referenced mosaic of an area of interest and classifies all the known elements (e.g., persons) found on the ground by an improved Faster R-CNN architecture previously trained. In subsequent reconnaissance flights, the system searches for any changes (e.g., disappearance of persons) that may occur in the mosaic by a histogram equalization and RGB-Local Binary Pattern (RGB-LBP) based algorithm. If present, the mosaic is updated. The second mode, allows to perform a real-time classification by using, again, our improved Faster R-CNN model, useful for time-critical operations. Thanks to different design features, the system works in real-time and performs mosaicking and change detection tasks at low-altitude, thus allowing the classification even of small objects. The proposed system was tested by using the whole set of challenging video sequences contained in the UAV Mosaicking and Change Detection (UMCD) dataset and other public datasets. The evaluation of the system by well-known performance metrics has shown remarkable results in terms of mosaic creation and updating, as well as in terms of change detection and object detection

    CGAMES'2009

    Get PDF

    Robust Localization in 3D Prior Maps for Autonomous Driving.

    Full text link
    In order to navigate autonomously, many self-driving vehicles require precise localization within an a priori known map that is annotated with exact lane locations, traffic signs, and additional metadata that govern the rules of the road. This approach transforms the extremely difficult and unpredictable task of online perception into a more structured localization problem—where exact localization in these maps provides the autonomous agent a wealth of knowledge for safe navigation. This thesis presents several novel localization algorithms that leverage a high-fidelity three-dimensional (3D) prior map that together provide a robust and reliable framework for vehicle localization. First, we present a generic probabilistic method for localizing an autonomous vehicle equipped with a 3D light detection and ranging (LIDAR) scanner. This proposed algorithm models the world as a mixture of several Gaussians, characterizing the z-height and reflectivity distribution of the environment—which we rasterize to facilitate fast and exact multiresolution inference. Second, we propose a visual localization strategy that replaces the expensive 3D LIDAR scanners with significantly cheaper, commodity cameras. In doing so, we exploit a graphics processing unit to generate synthetic views of our belief environment, resulting in a localization solution that achieves a similar order of magnitude error rate with a sensor that is several orders of magnitude cheaper. Finally, we propose a visual obstacle detection algorithm that leverages knowledge of our high-fidelity prior maps in its obstacle prediction model. This not only provides obstacle awareness at high rates for vehicle navigation, but also improves our visual localization quality as we are cognizant of static and non-static regions of the environment. All of these proposed algorithms are demonstrated to be real-time solutions for our self-driving car.PhDComputer Science and EngineeringUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttp://deepblue.lib.umich.edu/bitstream/2027.42/133410/1/rwolcott_1.pd

    Efficient Semantic Segmentation on Edge Devices

    Full text link
    Semantic segmentation works on the computer vision algorithm for assigning each pixel of an image into a class. The task of semantic segmentation should be performed with both accuracy and efficiency. Most of the existing deep FCNs yield to heavy computations and these networks are very power hungry, unsuitable for real-time applications on portable devices. This project analyzes current semantic segmentation models to explore the feasibility of applying these models for emergency response during catastrophic events. We compare the performance of real-time semantic segmentation models with non-real-time counterparts constrained by aerial images under oppositional settings. Furthermore, we train several models on the Flood-Net dataset, containing UAV images captured after Hurricane Harvey, and benchmark their execution on special classes such as flooded buildings vs. non-flooded buildings or flooded roads vs. non-flooded roads. In this project, we developed a real-time UNet based model and deployed that network on Jetson AGX Xavier module

    The XMM-Newton serendipitous ultraviolet source survey catalogue

    Get PDF
    The XMM-Newton Serendipitous Ultraviolet Source Survey (XMM-SUSS) is a catalogue of ultraviolet (UV) sources detected serendipitously by the Optical Monitor (XMM-OM) on-board the XMM-Newton observatory. The catalogue contains ultraviolet-detected sources collected from 2,417 XMM-OM observations in 1-6 broad band UV and optical filters, made between 24 February 2000 and 29 March 2007. The primary contents of the catalogue are source positions, magnitudes and fluxes in 1 to 6 passbands, and these are accompanied by profile diagnostics and variability statistics. The XMM-SUSS is populated by 753,578 UV source detections above a 3 sigma signal-to-noise threshold limit which relate to 624,049 unique objects. Taking account of substantial overlaps between observations, the net sky area covered is 29-54 square degrees, depending on UV filter. The magnitude distributions peak at 20.2, 20.9 and 21.2 in UVW2, UVM2 and UVW1 respectively. More than 10 per cent of sources have been visited more than once using the same filter during XMM-Newton operation, and > 20 per cent of sources are observed more than once per filter during an individual visit. Consequently, the scope for science based on temporal source variability on timescales of hours to years is broad. By comparison with other astrophysical catalogues we test the accuracy of the source measurements and define the nature of the serendipitous UV XMM-OM source sample. The distributions of source colours in the UV and optical filters are shown together with the expected loci of stars and galaxies, and indicate that sources which are detected in multiple UV bands are predominantly star-forming galaxies and stars of type G or earlier.Comment: Accepted for publication in MNRA

    Large-Scale Textured 3D Scene Reconstruction

    Get PDF
    Die Erstellung dreidimensionaler Umgebungsmodelle ist eine fundamentale Aufgabe im Bereich des maschinellen Sehens. Rekonstruktionen sind fĂŒr eine Reihe von Anwendungen von Nutzen, wie bei der Vermessung, dem Erhalt von KulturgĂŒtern oder der Erstellung virtueller Welten in der Unterhaltungsindustrie. Im Bereich des automatischen Fahrens helfen sie bei der BewĂ€ltigung einer Vielzahl an Herausforderungen. Dazu gehören Lokalisierung, das Annotieren großer DatensĂ€tze oder die vollautomatische Erstellung von Simulationsszenarien. Die Herausforderung bei der 3D Rekonstruktion ist die gemeinsame SchĂ€tzung von Sensorposen und einem Umgebunsmodell. Redundante und potenziell fehlerbehaftete Messungen verschiedener Sensoren mĂŒssen in eine gemeinsame ReprĂ€sentation der Welt integriert werden, um ein metrisch und photometrisch korrektes Modell zu erhalten. Gleichzeitig muss die Methode effizient Ressourcen nutzen, um Laufzeiten zu erreichen, welche die praktische Nutzung ermöglichen. In dieser Arbeit stellen wir ein Verfahren zur Rekonstruktion vor, das fĂ€hig ist, photorealistische 3D Rekonstruktionen großer Areale zu erstellen, die sich ĂŒber mehrere Kilometer erstrecken. Entfernungsmessungen aus Laserscannern und Stereokamerasystemen werden zusammen mit Hilfe eines volumetrischen Rekonstruktionsverfahrens fusioniert. RingschlĂŒsse werden erkannt und als zusĂ€tzliche Bedingungen eingebracht, um eine global konsistente Karte zu erhalten. Das resultierende Gitternetz wird aus Kamerabildern texturiert, wobei die einzelnen Beobachtungen mit ihrer GĂŒte gewichtet werden. FĂŒr eine nahtlose Erscheinung werden die unbekannten Belichtungszeiten und Parameter des optischen Systems mitgeschĂ€tzt und die Bilder entsprechend korrigiert. Wir evaluieren unsere Methode auf synthetischen Daten, realen Sensordaten unseres Versuchsfahrzeugs und öffentlich verfĂŒgbaren DatensĂ€tzen. Wir zeigen qualitative Ergebnisse großer innerstĂ€dtischer Bereiche, sowie quantitative Auswertungen der Fahrzeugtrajektorie und der RekonstruktionsqualitĂ€t. Zuletzt prĂ€sentieren wir mehrere Anwendungen und zeigen somit den Nutzen unserer Methode fĂŒr Anwendungen im Bereich des automatischen Fahrens

    Perception and intelligent localization for autonomous driving

    Get PDF
    Mestrado em Engenharia de Computadores e TelemĂĄticaVisĂŁo por computador e fusĂŁo sensorial sĂŁo temas relativamente recentes, no entanto largamente adoptados no desenvolvimento de robĂŽs autĂłnomos que exigem adaptabilidade ao seu ambiente envolvente. Esta dissertação foca-se numa abordagem a estes dois temas para alcançar percepção no contexto de condução autĂłnoma. O uso de cĂąmaras para atingir este fim Ă© um processo bastante complexo. Ao contrĂĄrio dos meios sensoriais clĂĄssicos que fornecem sempre o mesmo tipo de informação precisa e atingida de forma determinĂ­stica, as sucessivas imagens adquiridas por uma cĂąmara estĂŁo repletas da mais variada informação e toda esta ambĂ­gua e extremamente difĂ­cil de extrair. A utilização de cĂąmaras como meio sensorial em robĂłtica Ă© o mais prĂłximo que chegamos na semelhança com aquele que Ă© o de maior importĂąncia no processo de percepção humana, o sistema de visĂŁo. VisĂŁo por computador Ă© uma disciplina cientĂ­fica que engloba Ă reas como: processamento de sinal, inteligĂȘncia artificial, matemĂĄtica, teoria de controlo, neurobiologia e fĂ­sica. A plataforma de suporte ao estudo desenvolvido no Ăąmbito desta dissertação Ă© o ROTA (RObĂŽ Triciclo AutĂłnomo) e todos os elementos que consistem o seu ambiente. No contexto deste, sĂŁo descritas abordagens que foram introduzidas com fim de desenvolver soluçÔes para todos os desafios que o robĂŽ enfrenta no seu ambiente: detecção de linhas de estrada e consequente percepção desta, detecção de obstĂĄculos, semĂĄforos, zona da passadeira e zona de obras. É tambĂ©m descrito um sistema de calibração e aplicação da remoção da perspectiva da imagem, desenvolvido de modo a mapear os elementos percepcionados em distĂąncias reais. Em consequĂȘncia do sistema de percepção, Ă© ainda abordado o desenvolvimento de auto-localização integrado numa arquitectura distribuĂ­da incluindo navegação com planeamento inteligente. Todo o trabalho desenvolvido no decurso da dissertação Ă© essencialmente centrado no desenvolvimento de percepção robĂłtica no contexto de condução autĂłnoma.Computer vision and sensor fusion are subjects that are quite recent, however widely adopted in the development of autonomous robots that require adaptability to their surrounding environment. This thesis gives an approach on both in order to achieve perception in the scope of autonomous driving. The use of cameras to achieve this goal is a rather complex subject. Unlike the classic sensorial devices that provide the same type of information with precision and achieve this in a deterministic way, the successive images acquired by a camera are replete with the most varied information, that this ambiguous and extremely dificult to extract. The use of cameras for robotic sensing is the closest we got within the similarities with what is of most importance in the process of human perception, the vision system. Computer vision is a scientific discipline that encompasses areas such as signal processing, artificial intelligence, mathematics, control theory, neurobiology and physics. The support platform in which the study within this thesis was developed, includes ROTA (RObĂŽ Triciclo AutĂłnomo) and all elements comprising its environment. In its context, are described approaches that introduced in the platform in order to develop solutions for all the challenges facing the robot in its environment: detection of lane markings and its consequent perception, obstacle detection, trafic lights, crosswalk and road maintenance area. It is also described a calibration system and implementation for the removal of the image perspective, developed in order to map the elements perceived in actual real world distances. As a result of the perception system development, it is also addressed self-localization integrated in a distributed architecture that allows navigation with long term planning. All the work developed in the course of this work is essentially focused on robotic perception in the context of autonomous driving

    ReconViguRation: Reconfiguring Physical Keyboards in Virtual Reality.

    Get PDF
    Physical keyboards are common peripherals for personal computers and are efficient standard text entry devices. Recent research has investigated how physical keyboards can be used in immersive head-mounted display-based Virtual Reality (VR). So far, the physical layout of keyboards has typically been transplanted into VR for replicating typing experiences in a standard desktop environment. In this paper, we explore how to fully leverage the immersiveness of VR to change the input and output characteristics of physical keyboard interaction within a VR environment. This allows individual physical keys to be reconfigured to the same or different actions and visual output to be distributed in various ways across the VR representation of the keyboard. We explore a set of input and output mappings for reconfiguring the virtual presentation of physical keyboards and probe the resulting design space by specifically designing, implementing and evaluating nine VR-relevant applications: emojis, languages and special characters, application shortcuts, virtual text processing macros, a window manager, a photo browser, a whack-a-mole game, secure password entry and a virtual touch bar. We investigate the feasibility of the applications in a user study with 20 participants and find that, among other things, they are usable in VR. We discuss the limitations and possibilities of remapping the input and output characteristics of physical keyboards in VR based on empirical findings and analysis and suggest future research directions in this area
    • 

    corecore