31 research outputs found

    Generative Mesh Modeling

    Get PDF
    Generative Modeling is an alternative approach for the description of three-dimensional shape. The basic idea is to represent a model not as usual by an agglomeration of geometric primitives (triangles, point clouds, NURBS patches), but by functions. The paradigm change from objects to operations allows for a procedural representation of procedural shapes, such as most man-made objects. Instead of storing only the result of a 3D construction, the construction process itself is stored in a model file. The generative approach opens truly new perspectives in many ways, among others also for 3D knowledge management. It permits for instance to resort to a repository of already solved modeling problems, in order to re-use this knowledge also in different, slightly varied situations. The construction knowledge can be collected in digital libraries containing domain-specific parametric modeling tools. A concrete realization of this approach is a new general description language for 3D models, the "Generative Modeling Language" GML. As a Turing-complete "shape programming language" it is a basis of existing, primitv based 3D model formats. Together with its Runtime engine the GML permits - to store highly complex 3D models in a compact form, - to evaluate the description within fractions of a second, - to adaptively tesselate and to interactively display the model, - and even to change the models high-level parameters at runtime.Die generative Modellierung ist ein alternativer Ansatz zur Beschreibung von dreidimensionaler Form. Zugrunde liegt die Idee, ein Modell nicht wie üblich durch eine Ansammlung geometrischer Primitive (Dreiecke, Punkte, NURBS-Patches) zu beschreiben, sondern durch Funktionen. Der Paradigmenwechsel von Objekten zu Geometrie-erzeugenden Operationen ermöglicht es, prozedurale Modelle auch prozedural zu repräsentieren. Statt das Resultat eines 3D-Konstruktionsprozesses zu speichern, kann so der Konstruktionsprozess selber repräsentiert werden. Der generative Ansatz eröffnet unter anderem gänzlich neue Perspektiven für das Wissensmanagement im 3D-Bereich. Er ermöglicht etwa, auf einen Fundus bereits gelöster Konstruktions-Aufgaben zurückzugreifen, um sie in ähnlichen, aber leicht variierten Situationen wiederverwenden zu können. Das Konstruktions-Wissen kann dazu in Form von Bibliotheken parametrisierter, Domänen-spezifischer Modellier-Werkzeuge gesammelt werden. Konkret wird dazu eine neue allgemeine Modell-Beschreibungs-Sprache vorgeschlagen, die "Generative Modeling Language" GML. Als Turing-mächtige "Programmiersprache für Form" stellt sie eine echte Verallgemeinerung existierender Primitiv-basierter 3D-Modellformate dar. Zusammen mit ihrer Runtime-Engine erlaubt die GML, - hochkomplexe 3D-Objekte extrem kompakt zu beschreiben, - die Beschreibung innerhalb von Sekundenbruchteilen auszuwerten, - das Modell adaptiv darzustellen und interaktiv zu betrachten, - und die Modell-Parameter interaktiv zu verändern

    GPU data structures for graphics and vision

    Get PDF
    Graphics hardware has in recent years become increasingly programmable, and its programming APIs use the stream processor model to expose massive parallelization to the programmer. Unfortunately, the inherent restrictions of the stream processor model, used by the GPU in order to maintain high performance, often pose a problem in porting CPU algorithms for both video and volume processing to graphics hardware. Serial data dependencies which accelerate CPU processing are counterproductive for the data-parallel GPU. This thesis demonstrates new ways for tackling well-known problems of large scale video/volume analysis. In some instances, we enable processing on the restricted hardware model by re-introducing algorithms from early computer graphics research. On other occasions, we use newly discovered, hierarchical data structures to circumvent the random-access read/fixed write restriction that had previously kept sophisticated analysis algorithms from running solely on graphics hardware. For 3D processing, we apply known game graphics concepts such as mip-maps, projective texturing, and dependent texture lookups to show how video/volume processing can benefit algorithmically from being implemented in a graphics API. The novel GPU data structures provide drastically increased processing speed, and lift processing heavy operations to real-time performance levels, paving the way for new and interactive vision/graphics applications.Graphikhardware wurde in den letzen Jahren immer weiter programmierbar. Ihre APIs verwenden das Streamprozessor-Modell, um die massive Parallelisierung auch für den Programmierer verfügbar zu machen. Leider folgen aus dem strikten Streamprozessor-Modell, welches die GPU für ihre hohe Rechenleistung benötigt, auch Hindernisse in der Portierung von CPU-Algorithmen zur Video- und Volumenverarbeitung auf die GPU. Serielle Datenabhängigkeiten beschleunigen zwar CPU-Verarbeitung, sind aber für die daten-parallele GPU kontraproduktiv . Diese Arbeit präsentiert neue Herangehensweisen für bekannte Probleme der Video- und Volumensverarbeitung. Teilweise wird die Verarbeitung mit Hilfe von modifizierten Algorithmen aus der frühen Computergraphik-Forschung an das beschränkte Hardwaremodell angepasst. Anderswo helfen neu entdeckte, hierarchische Datenstrukturen beim Umgang mit den Schreibzugriff-Restriktionen die lange die Portierung von komplexeren Bildanalyseverfahren verhindert hatten. In der 3D-Verarbeitung nutzen wir bekannte Konzepte aus der Computerspielegraphik wie Mipmaps, projektive Texturierung, oder verkettete Texturzugriffe, und zeigen auf welche Vorteile die Video- und Volumenverarbeitung aus hardwarebeschleunigter Graphik-API-Implementation ziehen kann. Die präsentierten GPU-Datenstrukturen bieten drastisch schnellere Verarbeitung und heben rechenintensive Operationen auf Echtzeit-Niveau. Damit werden neue, interaktive Bildverarbeitungs- und Graphik-Anwendungen möglich

    A walk through the planned CS building

    Get PDF
    Using the architectural plan views of our future computer science building as test objects, we have completed the first stage of a Building walkthrough system. The inputs to our system are AutoCAD files. An AutoCAD converter translates the geometrical information in these files into a format suitable for 3D rendering. Major model errors, such as incorrect polygon intersections and random face orientations, are detected and fixed automatically. Interactive viewing and editing tools are provided to view the results, to modify and clean the model and to change surface attributes. Our display system provides a simple-to-use user interface for interactive exploration of buildings. Using only the mouse buttons, the user can move inside and outside the building and change floors. Several viewing and rendering options are provided, such as restricting the viewing frustum, avoiding wall collisions, and selecting different rendering algorithms. A plan view of the current floor, with the position of the eye point and viewing direction on it, is displayed at all times. The scene illumination can be manipulated, by interactively controlling intensity values for 5 light sources

    LocID - A Unique Object-at-a-Location Identifier; designing a global hierarchical geographical identifier that accounts for spatial inaccuracy and computational performance

    Get PDF
    Numerous organizations, such as insurance companies, work with global geospatial data. Global spatial operations, including risk assessment or disaster analysis, take time to compute. Spatial identifiers are a possible solution to speed up important spatial operations. Modern spatial identifiers struggle to encode geometries: Distortion of area, found in the common Mercator projection, poses challenges for precise area calculations and comparisons across latitudes. Complex transformation of areas into local Cartesian coordinate systems is thus needed. The added complexity increases errors. Latitude and longitude coordinates, while precise for location, lack the ability to convey object size or globally consistent location accuracy. Thus, a hierarchical system as outlined in this thesis may well be a solution for performant and accurate object identification. Discrete global grid systems (DGGSs), offer global continuous indexing, maintain consistent spatial resolution, and support data aggregation. This thesis compares different DGGS options based on factors like hierarchical structure, tesselation shape, accuracy, and programming language support. This work demonstrates that triangles, as a tesselation shape, provide consistent area sizes and computational efficiency, making them suitable for LocID’s goal of addressing spatial challenges in risk assessment and disaster analysis on the Earth’s surface. Consequently, the triangle-based DGGS Quarternary Triangular Mesh (QTM) is selected as the foundation for building LocID. The goal of LocID is to identify, match, recognize contains, and detect changes in objects using only its ID attribute. The process of designing, conceptualizing, and building a reference implementation of LocID follows a circular development loop, where requirements are defined, researched, designed, implemented, and evaluated iteratively. The encoding uses Quaternary Triangular Mesh (QTM), a DGGS based on an octahedron and on triangles as its cell shape. LocID identifiers consist of encoding digits and separators. LocID’s geometry part is encoded recursively, with the number of digits indicating the number of levels used for encoding an object. This part is also compressed in a tree-like manner, by not double-encoding branches shared by leaves, leaves being the highest-level QTM cells used for encoding. A reference Python implementation, called LocID.py, is provided. Examples and tests are included to verify the correctness of encoding functions and operations. By encoding geometries within Discrete Global Grid System (DGGS) cells, LocID enables spatial operations without the need for additional spatial computation, providing quick string or byte operations such as matching and containment. LocID extends a DGGS, enabling geometry encoding and processing within cells. DGGSs offer a path to more performant geospatial data processing for points, whereas geometries have not yet been considered. LocID addresses this, offering a solution for encoding geometries in a way that allows a few operations to be performed performantly directly on the ID itself

    New Models for High-Quality Surface Reconstruction and Rendering

    Get PDF
    The efficient reconstruction and artifact-free visualization of surfaces from measured real-world data is an important issue in various applications, such as medical and scientific visualization, quality control, and the media-related industry. The main contribution of this thesis is the development of the first efficient GPU-based reconstruction and visualization methods using trivariate splines, i.e., splines defined on tetrahedral partitions. Our methods show that these models are very well-suited for real-time reconstruction and high-quality visualizations of surfaces from volume data. We create a new quasi-interpolating operator which for the first time solves the problem of finding a globally C1-smooth quadratic spline approximating data and where no tetrahedra need to be further subdivided. In addition, we devise a new projection method for point sets arising from a sufficiently dense sampling of objects. Compared with existing approaches, high-quality surface triangulations can be generated with guaranteed numerical stability. Keywords. Piecewise polynomials; trivariate splines; quasi-interpolation; volume data; GPU ray casting; surface reconstruction; point set surface

    New Models for High-Quality Surface Reconstruction and Rendering

    Get PDF
    The efficient reconstruction and artifact-free visualization of surfaces from measured real-world data is an important issue in various applications, such as medical and scientific visualization, quality control, and the media-related industry. The main contribution of this thesis is the development of the first efficient GPU-based reconstruction and visualization methods using trivariate splines, i.e., splines defined on tetrahedral partitions. Our methods show that these models are very well-suited for real-time reconstruction and high-quality visualizations of surfaces from volume data. We create a new quasi-interpolating operator which for the first time solves the problem of finding a globally C1-smooth quadratic spline approximating data and where no tetrahedra need to be further subdivided. In addition, we devise a new projection method for point sets arising from a sufficiently dense sampling of objects. Compared with existing approaches, high-quality surface triangulations can be generated with guaranteed numerical stability. Keywords. Piecewise polynomials; trivariate splines; quasi-interpolation; volume data; GPU ray casting; surface reconstruction; point set surface

    Creating 3D models of cultural heritage sites with terrestrial laser scanning and 3D imaging

    Get PDF
    Includes abstract.Includes bibliographical references.The advent of terrestrial laser-scanners made the digital preservation of cultural heritage sites an affordable technique to produce accurate and detailed 3D-computermodel representations for any kind of 3D-objects, such as buildings, infrastructure, and even entire landscapes. However, one of the key issues with this technique is the large amount of recorded points; a problem which was even more intensified by the recent advances in laser-scanning technology, which increased the data acquisition rate from 25 thousand to 1 million points per second. The following research presents a workflow for the processing of large-volume laser-scanning data, with a special focus on the needs of the Zamani initiative. The research project, based at the University of Cape Town, spatially documents African Cultural Heritage sites and Landscapes and produces meshed 3D models, of various, historically important objects, such as fortresses, mosques, churches, castles, palaces, rock art shelters, statues, stelae and even landscapes

    Realistic Visualization of Accessories within Interactive Simulation Systems for Garment Prototyping

    Get PDF
    In virtual garment prototyping, designers create a garment design by using Computer Aided Design (CAD). In difference to traditional CAD the word "aided" in this case refers to the computer replicating real world behavior of garments. This allows the designer to interact naturally with his design. The designer has a wide range of expressions within his work. This is done by defining details on a garment which are not limited to the type of cloth used. The way how cloth patterns are sewn together and the style and usage of details of the cloth's surface, like appliqués, have a strong impact on the visual appearance of a garment to a large degree. Therefore, virtual and real garments usually have a lot of such surface details. Interactive virtual garment prototyping itself is an interdisciplinary field. Several problems have to be solved to create an efficiently usable real-time virtual prototyping system for garment manufacturers. Such a system can be roughly separated into three sub-components. The first component deals with acquisition of material and other data needed to let a simulation mimic plausible real world behavior of the garment. The second component is the garment simulation process itself. Finally, the third component is centered on the visualization of the simulation results. Therefore, the overall process spans several scientific areas which have to take into account the needs of each other in order to get an overall interactive system. In my work I especially target the third section, which deals with the visualization. On the scientific side, the developments in the last years have shown great improvements on both speed and reliability of simulation and rendering approaches suitable for the virtual prototyping of garments. However, with the currently existing approaches there are still many problems to be solved, especially if interactive simulation and visualization need to work together and many object and surface details come into play. This is the case when using a virtual prototyping in a productive environment. The currently available approaches try to handle most of the surface details as part of the simulation. This generates a lot of data early in the pipeline which needs to be transferred and processed, requiring a lot of processing time and easily stalls the pipeline defined by the simulation and visualization system. Additionally, real world garment examples are already complicated in their cloth arrangement alone. This requires additional computational power. Therefore, the interactive garment simulation tends to lose its capability to allow interactive handling of the garment. In my work I present a solution, which solves this problem by moving the handling of design details from the simulation stage entirely to a completely GPU based rendering stage. This way, the behavior of the garment and its visual appearance are separated. Therefore, the simulation part can fully concentrate on simulating the fabric behavior, while the visualization handles the placing of surface details lighting, materials and self-shadowing. Thus, a much higher degree of surface complexity can be achieved within an interactive virtual prototyping system as can be done with the current existing approaches

    Fully anisotropic split-tree adaptive refinement mesh generation using tetrahedral mesh stitching

    Get PDF
    Due to the myriad of geometric topologies that modern computational fluid dynamicists desire to mesh and run solutions on, the need for a robust Cartesian Mesh Generation algorithm is paramount. Not only do Cartesian meshes require less elements and often help resolve flow features but they also allow the grid generator to have a great deal of control in so far as element aspect ratio, size, and gradation. Fully Anisotropic Split-Tree Adaptive Refinement (FASTAR) is a code that allows the user to exert a great deal of control and ultimately generate a valid, geometry conforming mesh. Due to the split-tree nature and the use of volumetric pixels (voxels), non-unit aspect ratio meshing is easily achieved. Nodes are not generated until the end which mitigates tolerance issues. The tree is retained coherently, and viscous layers may be inserted in the space between the geometry and the Cartesian mesh before it is tetrahedralized. FASTAR uses tree traversal to determine neighbors robustly, and with the tetrahedralization of only a small amount of space around the geometry, sliver cells and inverted elements are avoided. The code uses Riemannian Metric Tensors to generate geometry-appropriate spacing and is capable of adaptive meshing from a spacing field generated either by the user or from solution data. FASTAR is a robust, general mesh generator that allows maximum flexibility with minimal post-processing
    corecore