158,881 research outputs found
On applications of chimera grid schemes to store separation
A finite difference scheme which uses multiple overset meshes to simulate the aerodynamics of aircraft/store interaction and store separation is described. In this chimera, or multiple mesh, scheme, a complex configuration is mapped using a major grid about the main component of the configuration, and minor overset meshes are used to map each additional component such as a store. As a first step in modeling the aerodynamics of store separation, two dimensional inviscid flow calculations were carried out in which one of the minor meshes is allowed to move with respect to the major grid. Solutions of calibrated two dimensional problems indicate that allowing one mesh to move with respect to another does not adversely affect the time accuracy of an unsteady solution. Steady, inviscid three dimensional computations demonstrate the capability to simulate complex configurations, including closely packed multiple bodies
Heap Abstractions for Static Analysis
Heap data is potentially unbounded and seemingly arbitrary. As a consequence,
unlike stack and static memory, heap memory cannot be abstracted directly in
terms of a fixed set of source variable names appearing in the program being
analysed. This makes it an interesting topic of study and there is an abundance
of literature employing heap abstractions. Although most studies have addressed
similar concerns, their formulations and formalisms often seem dissimilar and
some times even unrelated. Thus, the insights gained in one description of heap
abstraction may not directly carry over to some other description. This survey
is a result of our quest for a unifying theme in the existing descriptions of
heap abstractions. In particular, our interest lies in the abstractions and not
in the algorithms that construct them.
In our search of a unified theme, we view a heap abstraction as consisting of
two features: a heap model to represent the heap memory and a summarization
technique for bounding the heap representation. We classify the models as
storeless, store based, and hybrid. We describe various summarization
techniques based on k-limiting, allocation sites, patterns, variables, other
generic instrumentation predicates, and higher-order logics. This approach
allows us to compare the insights of a large number of seemingly dissimilar
heap abstractions and also paves way for creating new abstractions by
mix-and-match of models and summarization techniques.Comment: 49 pages, 20 figure
The Process of Generating Single Large Combined Cloud for Grid-Free Solvers
Dealing with moving body problems, where one component moves relatively with respect to other, is a difficult task in CFD due to the efforts needed in grid handling for every delta change in position of the moving component. The inherent nature of mesh-free solvers reduces the efforts needed for these kinds of problems by operating on a cloud of points rather than a grid. A new method to handle moving body problems is proposed, where individual clouds are generated around each component and are combined into a single large combined cloud. The proposed method is applied to store separation problem and results generated using NAL-MCIR mesh-less solver is compared with experimental results
An introduction to Graph Data Management
A graph database is a database where the data structures for the schema
and/or instances are modeled as a (labeled)(directed) graph or generalizations
of it, and where querying is expressed by graph-oriented operations and type
constructors. In this article we present the basic notions of graph databases,
give an historical overview of its main development, and study the main current
systems that implement them
Simulation of reaction-diffusion processes in three dimensions using CUDA
Numerical solution of reaction-diffusion equations in three dimensions is one
of the most challenging applied mathematical problems. Since these simulations
are very time consuming, any ideas and strategies aiming at the reduction of
CPU time are important topics of research. A general and robust idea is the
parallelization of source codes/programs. Recently, the technological
development of graphics hardware created a possibility to use desktop video
cards to solve numerically intensive problems. We present a powerful parallel
computing framework to solve reaction-diffusion equations numerically using the
Graphics Processing Units (GPUs) with CUDA. Four different reaction-diffusion
problems, (i) diffusion of chemically inert compound, (ii) Turing pattern
formation, (iii) phase separation in the wake of a moving diffusion front and
(iv) air pollution dispersion were solved, and additionally both the Shared
method and the Moving Tiles method were tested. Our results show that parallel
implementation achieves typical acceleration values in the order of 5-40 times
compared to CPU using a single-threaded implementation on a 2.8 GHz desktop
computer.Comment: 8 figures, 5 table
Polly's Polyhedral Scheduling in the Presence of Reductions
The polyhedral model provides a powerful mathematical abstraction to enable
effective optimization of loop nests with respect to a given optimization goal,
e.g., exploiting parallelism. Unexploited reduction properties are a frequent
reason for polyhedral optimizers to assume parallelism prohibiting dependences.
To our knowledge, no polyhedral loop optimizer available in any production
compiler provides support for reductions. In this paper, we show that
leveraging the parallelism of reductions can lead to a significant performance
increase. We give a precise, dependence based, definition of reductions and
discuss ways to extend polyhedral optimization to exploit the associativity and
commutativity of reduction computations. We have implemented a
reduction-enabled scheduling approach in the Polly polyhedral optimizer and
evaluate it on the standard Polybench 3.2 benchmark suite. We were able to
detect and model all 52 arithmetic reductions and achieve speedups up to
2.21 on a quad core machine by exploiting the multidimensional
reduction in the BiCG benchmark.Comment: Presented at the IMPACT15 worksho
Background Subtraction via Fast Robust Matrix Completion
Background subtraction is the primary task of the majority of video
inspection systems. The most important part of the background subtraction which
is common among different algorithms is background modeling. In this regard,
our paper addresses the problem of background modeling in a computationally
efficient way, which is important for current eruption of "big data" processing
coming from high resolution multi-channel videos. Our model is based on the
assumption that background in natural images lies on a low-dimensional
subspace. We formulated and solved this problem in a low-rank matrix completion
framework. In modeling the background, we benefited from the in-face extended
Frank-Wolfe algorithm for solving a defined convex optimization problem. We
evaluated our fast robust matrix completion (fRMC) method on both background
models challenge (BMC) and Stuttgart artificial background subtraction (SABS)
datasets. The results were compared with the robust principle component
analysis (RPCA) and low-rank robust matrix completion (RMC) methods, both
solved by inexact augmented Lagrangian multiplier (IALM). The results showed
faster computation, at least twice as when IALM solver is used, while having a
comparable accuracy even better in some challenges, in subtracting the
backgrounds in order to detect moving objects in the scene
A formal semantics for control and data flow in the gannet service-based system-on-chip architecture
There is a growing demand for solutions which allow the design of large and complex reconfigurable Systems-on-
Chip (SoC) at high abstraction levels. The Gannet project proposes a functional programming approach for high-abstraction design of very large SoCs. Gannet is a distributed service-based SoC architecture, i.e. a network of services offered by hardware or software cores. The Gannet SoC is task-level reconfigurable: it performs tasks by executing functional task description programs using a demand-driven dataflow mechanism. The Gannet architecture
combines the flexible connectivity offered by a Networkon-
Chip with the functional language paradigm to create a
fully concurrent distributed SoC with the option to completely separate data flows from control flows. This feature is essential to avoid a bottleneck at he controller for run-time control of multiple high-throughput data flows.
In this paper we present the Gannet architecture and language
and introduce an operational semantics to formally describe the mechanism to separate control and data flows
- …
