8,441 research outputs found

    Solving Vertical Transport and Chemistry in Air Pollution Models.

    Get PDF
    For the time integration of stiff transport-chemistry problems from air pollution modelling, standard ODE solvers are not feasible due to the large number of species and the 3D nature. The popular alternative, standard operator splitting, introduces artificial transients for short-lived species. This complicates the chemistry solution, easily causing large errors for such species. In the framework of an operational global air pollution model, we focus on the problem formed by chemistry and vertical transport, which is based on diffusion, cloud-related vertical winds, and wet deposition. Its specific nature leads to full Jacobian matrices, ruling out standard implicit integration. We compare Strang operator splitting with two alternatives: source splitting and an (unsplit) Rosenbrock method with approximate matrix factorization, all having equal computational cost. The comparison is performed with real data. All methods are applied with half-hour time steps, and give good accuracies. Rosenbrock is the most accurate, and source splitting is more accurate than Strang splitting. Splitting errors concentrate in short-lived species sensitive to solar radiation and species with strong emissions and depositions

    Simulation of reaction-diffusion processes in three dimensions using CUDA

    Get PDF
    Numerical solution of reaction-diffusion equations in three dimensions is one of the most challenging applied mathematical problems. Since these simulations are very time consuming, any ideas and strategies aiming at the reduction of CPU time are important topics of research. A general and robust idea is the parallelization of source codes/programs. Recently, the technological development of graphics hardware created a possibility to use desktop video cards to solve numerically intensive problems. We present a powerful parallel computing framework to solve reaction-diffusion equations numerically using the Graphics Processing Units (GPUs) with CUDA. Four different reaction-diffusion problems, (i) diffusion of chemically inert compound, (ii) Turing pattern formation, (iii) phase separation in the wake of a moving diffusion front and (iv) air pollution dispersion were solved, and additionally both the Shared method and the Moving Tiles method were tested. Our results show that parallel implementation achieves typical acceleration values in the order of 5-40 times compared to CPU using a single-threaded implementation on a 2.8 GHz desktop computer.Comment: 8 figures, 5 table

    An inverse Gaussian plume approach for estimating atmospheric pollutant emissions from multiple point sources

    Full text link
    A method is developed for estimating the emission rates of contaminants into the atmosphere from multiple point sources using measurements of particulate material deposited at ground level. The approach is based on a Gaussian plume type solution for the advection-diffusion equation with ground-level deposition and given emission sources. This solution to the forward problem is incorporated into an inverse algorithm for estimating the emission rates by means of a linear least squares approach. The results are validated using measured deposition and meteorological data from a large lead-zinc smelting operation in Trail, British Columbia. The algorithm is demonstrated to be robust and capable of generating reasonably accurate estimates of total contaminant emissions over the relatively short distances of interest in this study

    A new approximate matrix factorization for implicit time integration in air pollution modeling

    Get PDF
    Implicit time stepping typically requires solution of one or several linear systems with a matrix I−τJ per time step where J is the Jacobian matrix. If solution of these systems is expensive, replacing I−τJ with its approximate matrix factorization (AMF) (I−τR)(I−τV), R+V=J, often leads to a good compromise between stability and accuracy of the time integration on the one hand and its efficiency on the other hand. For example, in air pollution modeling, AMF has been successfully used in the framework of Rosenbrock schemes. The standard AMF gives an approximation to I−τJ with the error τ2RV, which can be significant in norm. In this paper we propose a new AMF. In assumption that −V is an M-matrix, the error of the new AMF can be shown to have an upper bound τ||R||, while still being asymptotically O(τ2)O(\tau^2). This new AMF, called AMF+, is equal in costs to standard AMF and, as both analysis and numerical experiments reveal, provides a better accuracy. We also report on our experience with another, cheaper AMF and with AMF-preconditioned GMRES

    A Reconfigurable Vector Instruction Processor for Accelerating a Convection Parametrization Model on FPGAs

    Full text link
    High Performance Computing (HPC) platforms allow scientists to model computationally intensive algorithms. HPC clusters increasingly use General-Purpose Graphics Processing Units (GPGPUs) as accelerators; FPGAs provide an attractive alternative to GPGPUs for use as co-processors, but they are still far from being mainstream due to a number of challenges faced when using FPGA-based platforms. Our research aims to make FPGA-based high performance computing more accessible to the scientific community. In this work we present the results of investigating the acceleration of a particular atmospheric model, Flexpart, on FPGAs. We focus on accelerating the most computationally intensive kernel from this model. The key contribution of our work is the architectural exploration we undertook to arrive at a solution that best exploits the parallelism available in the legacy code, and is also convenient to program, so that eventually the compilation of high-level legacy code to our architecture can be fully automated. We present the three different types of architecture, comparing their resource utilization and performance, and propose that an architecture where there are a number of computational cores, each built along the lines of a vector instruction processor, works best in this particular scenario, and is a promising candidate for a generic FPGA-based platform for scientific computation. We also present the results of experiments done with various configuration parameters of the proposed architecture, to show its utility in adapting to a range of scientific applications.Comment: This is an extended pre-print version of work that was presented at the international symposium on Highly Efficient Accelerators and Reconfigurable Technologies (HEART2014), Sendai, Japan, June 911, 201

    PARALLEL COMPUTATIONS WITH LARGE-SCALE AIR\ud POLLUTION MODELS

    Get PDF
    Large-scale mathematical models are very powerful tools in the efforts to provide more\ud information and more detailed information about the pollution levels, especially about pollution\ud levels which exceed certain critical values.. However, the model used must satisfy at\ud least two conditions: (i) it must be verified that the model results are reliable and (ii) it\ud should be possible to carry out different study by using the model. It is clear that comprehensive\ud studies about relationships between different input parameters and the model results\ud can only be carried out (a) if the numerical methods used in the model are sufficiently\ud fast and (b) if the code runs efficiently on the available high-speed computers.\ud Some results obtained recently by a new unified version of the Danish Eulerian Model will\ud be presented in this paper

    Coupling traffic models on networks and urban dispersion models for simulating sustainable mobility strategies

    Get PDF
    AbstractThe aim of the present paper is to investigate the viability of macroscopic traffic models for modeling and testing different traffic scenarios, in order to define the impact on air quality of different strategies for the reduction of traffic emissions. To this aim, we complement a well assessed traffic model on networks (Garavello and Piccoli (2006) [1]) with a strategy for estimating data needed from the model and we couple it with the urban dispersion model Sirane (Soulhac (2000) [2])

    Fast stable direct fitting and smoothness selection for Generalized Additive Models

    Get PDF
    Existing computationally efficient methods for penalized likelihood GAM fitting employ iterative smoothness selection on working linear models (or working mixed models). Such schemes fail to converge for a non-negligible proportion of models, with failure being particularly frequent in the presence of concurvity. If smoothness selection is performed by optimizing `whole model' criteria these problems disappear, but until now attempts to do this have employed finite difference based optimization schemes which are computationally inefficient, and can suffer from false convergence. This paper develops the first computationally efficient method for direct GAM smoothness selection. It is highly stable, but by careful structuring achieves a computational efficiency that leads, in simulations, to lower mean computation times than the schemes based on working-model smoothness selection. The method also offers a reliable way of fitting generalized additive mixed models
    corecore