503 research outputs found
Recursive Algorithms for Distributed Forests of Octrees
The forest-of-octrees approach to parallel adaptive mesh refinement and
coarsening (AMR) has recently been demonstrated in the context of a number of
large-scale PDE-based applications. Although linear octrees, which store only
leaf octants, have an underlying tree structure by definition, it is not often
exploited in previously published mesh-related algorithms. This is because the
branches are not explicitly stored, and because the topological relationships
in meshes, such as the adjacency between cells, introduce dependencies that do
not respect the octree hierarchy. In this work we combine hierarchical and
topological relationships between octree branches to design efficient recursive
algorithms.
We present three important algorithms with recursive implementations. The
first is a parallel search for leaves matching any of a set of multiple search
criteria. The second is a ghost layer construction algorithm that handles
arbitrarily refined octrees that are not covered by previous algorithms, which
require a 2:1 condition between neighboring leaves. The third is a universal
mesh topology iterator. This iterator visits every cell in a domain partition,
as well as every interface (face, edge and corner) between these cells. The
iterator calculates the local topological information for every interface that
it visits, taking into account the nonconforming interfaces that increase the
complexity of describing the local topology. To demonstrate the utility of the
topology iterator, we use it to compute the numbering and encoding of
higher-order nodal basis functions.
We analyze the complexity of the new recursive algorithms theoretically, and
assess their performance, both in terms of single-processor efficiency and in
terms of parallel scalability, demonstrating good weak and strong scaling up to
458k cores of the JUQUEEN supercomputer.Comment: 35 pages, 15 figures, 3 table
Algorithms and methods for discrete mesh repair
Computational analysis and design has become a fundamental part of product research, development, and manufacture in aerospace, automotive, and other industries. In general the success of the specific application depends heavily on the accuracy and consistency of the computational model used. The aim of this work is to reduce the time needed to prepare geometry for mesh generation. This will be accomplished by developing tools that semi-automatically repair discrete data. Providing a level of automation to the process of repairing large, complex problems in discrete data will significantly accelerate the grid generation process. The developed algorithms are meant to offer semi-automated solutions to complicated geometrical problems—specifically discrete mesh intersections and isolated boundaries. The intersection-repair strategy presented here focuses on repairing the intersection in-place as opposed to re-discretizing the intersecting geometries. Combining robust, efficient methods of detecting intersections and then repairing intersecting geometries in-place produces a significant improvement over techniques used in current literature. The result of this intersection process is a non-manifold, non-intersecting geometry that is free of duplicate and degenerate geometry. Results are presented showing the accuracy and consistency of the intersection repair tool. Isolated boundaries are a type of gap that current research does not address directly. They are defined by discrete boundary edges that are unable to be paired with nearby discrete boundary edges in order to fill the existing gap. In this research the method of repair seeks to fill the gap by extruding the isolated boundary along a defined vector so that it is topologically adjacent to a nearby surface. The outcome of the repair process is that the isolated boundaries no longer exist because the gap has been filled. Results are presented showing the precision of the edge projection and the advantage of edge splitting in the repair of isolated boundaries
Sensory processing and world modeling for an active ranging device
In this project, we studied world modeling and sensory processing for laser range data. World Model data representation and operation were defined. Sensory processing algorithms for point processing and linear feature detection were designed and implemented. The interface between world modeling and sensory processing in the Servo and Primitive levels was investigated and implemented. In the primitive level, linear features detectors for edges were also implemented, analyzed and compared. The existing world model representations is surveyed. Also presented is the design and implementation of the Y-frame model, a hierarchical world model. The interfaces between the world model module and the sensory processing module are discussed as well as the linear feature detectors that were designed and implemented
High-Quality Simplification and Repair of Polygonal Models
Because of the rapid evolution of 3D acquisition and modelling methods, highly complex and detailed polygonal models with constantly increasing polygon count are used as three-dimensional geometric representations of objects in computer graphics and engineering applications. The fact that this particular representation is arguably the most widespread one is due to its simplicity, flexibility and rendering support by 3D graphics hardware. Polygonal models are used for rendering of objects in a broad range of disciplines like medical imaging, scientific visualization, computer aided design, film industry, etc. The handling of huge scenes composed of these high-resolution models rapidly approaches the computational capabilities of any graphics accelerator. In order to be able to cope with the complexity and to build level-of-detail representations, concentrated efforts were dedicated in the recent years to the development of new mesh simplification methods that produce high-quality approximations of complex models by reducing the number of polygons used in the surface while keeping the overall shape, volume and boundaries preserved as much as possible. Many well-established methods and applications require "well-behaved" models as input. Degenerate or incorectly oriented faces, T-joints, cracks and holes are just a few of the possible degenaracies that are often disallowed by various algorithms. Unfortunately, it is all too common to find polygonal models that contain, due to incorrect modelling or acquisition, such artefacts. Applications that may require "clean" models include finite element analysis, surface smoothing, model simplification, stereo lithography. Mesh repair is the task of removing artefacts from a polygonal model in order to produce an output model that is suitable for further processing by methods and applications that have certain quality requirements on their input. This thesis introduces a set of new algorithms that address several particular aspects of mesh repair and mesh simplification. One of the two mesh repair methods is dealing with the inconsistency of normal orientation, while another one, removes the inconsistency of vertex connectivity. Of the three mesh simplification approaches presented here, the first one attempts to simplify polygonal models with the highest possible quality, the second, applies the developed technique to out-of-core simplification, and the third, prevents self-intersections of the model surface that can occur during mesh simplification
Scalable Adaptive Mantle Convection Simulation on Petascale Supercomputers
Mantle convection is the principal control on
the thermal and geological evolution of the Earth. Mantle
convection modeling involves solution of the mass, momentum,
and energy equations for a viscous, creeping, incompressible
non-Newtonian fluid at high Rayleigh and Peclet
numbers. Our goal is to conduct global mantle convection
simulations that can resolve faulted plate boundaries, down
to 1 km scales. However, uniform resolution at these scales
would result in meshes with a trillion elements, which
would elude even sustained petaflops supercomputers. Thus
parallel adaptive mesh refinement and coarsening (AMR)
is essential.
We present RHEA, a new generation mantle convection
code designed to scale to hundreds of thousands of cores.
RHEA is built on ALPS, a parallel octree-based adaptive
mesh finite element library that provides new distributed
data structures and parallel algorithms for dynamic coarsening,
refinement, rebalancing, and repartitioning of the
mesh. ALPS currently supports low order continuous
Lagrange elements, and arbitrary order discontinuous
Galerkin spectral elements, on octree meshes. A forest-ofoctrees
implementation permits nearly arbitrary geometries
to be accommodated. Using TACC’s 579 teraflops
Ranger supercomputer, we demonstrate excellent weak and
strong scalability of parallel AMR on up to 62,464 cores
for problems with up to 12.4 billion elements. With RHEA’s
adaptive capabilities, we have been able to reduce the
number of elements by over three orders of magnitude,
thus enabling us to simulate large-scale mantle convection
with finest local resolution of 1.5 km
- …