9,598 research outputs found

    Space Vector Modulation Techniques for Multilevel Converters – a survey

    Get PDF
    This paper presents a survey of most recent, simple and efficient Space Vector Modulation algorithms for multilevel converters. These algorithms avoid trigonometric and other complex operations, leading to more simple and cost efficient implementations. They can be applied to multilevel topologies and present freedom degrees that can be Exploited in order to optimize system parameters in the system like: capacitors voltages balancing or voltage/current ripples. Experimental results are presented to show the good performance of the algorithms

    MAGMA: Multi-level accelerated gradient mirror descent algorithm for large-scale convex composite minimization

    Full text link
    Composite convex optimization models arise in several applications, and are especially prevalent in inverse problems with a sparsity inducing norm and in general convex optimization with simple constraints. The most widely used algorithms for convex composite models are accelerated first order methods, however they can take a large number of iterations to compute an acceptable solution for large-scale problems. In this paper we propose to speed up first order methods by taking advantage of the structure present in many applications and in image processing in particular. Our method is based on multi-level optimization methods and exploits the fact that many applications that give rise to large scale models can be modelled using varying degrees of fidelity. We use Nesterov's acceleration techniques together with the multi-level approach to achieve O(1/ϵ)\mathcal{O}(1/\sqrt{\epsilon}) convergence rate, where ϵ\epsilon denotes the desired accuracy. The proposed method has a better convergence rate than any other existing multi-level method for convex problems, and in addition has the same rate as accelerated methods, which is known to be optimal for first-order methods. Moreover, as our numerical experiments show, on large-scale face recognition problems our algorithm is several times faster than the state of the art

    HP-multigrid as smoother algorithm for higher order discontinuous Galerkin discretizations of advection dominated flows. Part II. Optimization of the Runge-Kutta smoother

    Get PDF
    Using a detailed multilevel analysis of the complete hp-Multigrid as Smoother algorithm accurate predictions are obtained of the spectral radius and operator norms of the multigrid error transformation operator. This multilevel analysis is used to optimize the coefficients in the semi-implicit Runge-Kutta smoother, such that the spectral radius of the multigrid error transformation operator is minimal under properly chosen constraints. The Runge-Kutta coefficients for a wide range of cell Reynolds numbers and a detailed analysis of the performance of the hp-MGS algorithm are presented. In addition, the computational complexity of the hp-MGS algorithm is investigated. The hp-MGS algorithm is tested on a fourth order accurate space-time discontinuous Galerkin finite element discretization of the advection-diffusion equation for a number of model problems, which include thin boundary layers and highly stretched meshes, and a non-constant advection velocity. For all test cases excellent multigrid convergence is obtained

    Relaxation-Based Coarsening for Multilevel Hypergraph Partitioning

    Get PDF
    Multilevel partitioning methods that are inspired by principles of multiscaling are the most powerful practical hypergraph partitioning solvers. Hypergraph partitioning has many applications in disciplines ranging from scientific computing to data science. In this paper we introduce the concept of algebraic distance on hypergraphs and demonstrate its use as an algorithmic component in the coarsening stage of multilevel hypergraph partitioning solvers. The algebraic distance is a vertex distance measure that extends hyperedge weights for capturing the local connectivity of vertices which is critical for hypergraph coarsening schemes. The practical effectiveness of the proposed measure and corresponding coarsening scheme is demonstrated through extensive computational experiments on a diverse set of problems. Finally, we propose a benchmark of hypergraph partitioning problems to compare the quality of other solvers

    Optimizing Omega

    Get PDF
    "The original publication is available at www.springerlink.com " Copyright Springer. DOI: 10.1007/s10898-008-9396-5This paper considers the Omega function, proposed by Cascon, Keating & Shadwick as a performance measure for comparing financial assets. We discuss the use of Omega as a basis for portfolio selection. We show that the problem of choosing portfolio weights in order to maximize Omega typically has many local solutions and we describe some preliminary computational experience of finding the global optimum using a NAG library implementation of the Huyer & Neumaier MCS method.Peer reviewe
    corecore