10 research outputs found

    A Parallel Approximation Algorithm for Positive Semidefinite Programming

    Full text link
    Positive semidefinite programs are an important subclass of semidefinite programs in which all matrices involved in the specification of the problem are positive semidefinite and all scalars involved are non-negative. We present a parallel algorithm, which given an instance of a positive semidefinite program of size N and an approximation factor eps > 0, runs in (parallel) time poly(1/eps) \cdot polylog(N), using poly(N) processors, and outputs a value which is within multiplicative factor of (1 + eps) to the optimal. Our result generalizes analogous result of Luby and Nisan [1993] for positive linear programs and our algorithm is inspired by their algorithm.Comment: 16 page

    Using Optimization to Obtain a Width-Independent, Parallel, Simpler, and Faster Positive SDP Solver

    Full text link
    We study the design of polylogarithmic depth algorithms for approximately solving packing and covering semidefinite programs (or positive SDPs for short). This is a natural SDP generalization of the well-studied positive LP problem. Although positive LPs can be solved in polylogarithmic depth while using only O~(log2n/ε2)\tilde{O}(\log^{2} n/\varepsilon^2) parallelizable iterations, the best known positive SDP solvers due to Jain and Yao require O(log14n/ε13)O(\log^{14} n /\varepsilon^{13}) parallelizable iterations. Several alternative solvers have been proposed to reduce the exponents in the number of iterations. However, the correctness of the convergence analyses in these works has been called into question, as they both rely on algebraic monotonicity properties that do not generalize to matrix algebra. In this paper, we propose a very simple algorithm based on the optimization framework proposed for LP solvers. Our algorithm only needs O~(log2n/ε2)\tilde{O}(\log^2 n / \varepsilon^2) iterations, matching that of the best LP solver. To surmount the obstacles encountered by previous approaches, our analysis requires a new matrix inequality that extends Lieb-Thirring's inequality, and a sign-consistent, randomized variant of the gradient truncation technique proposed in

    A novel way to automatically plan cellular networks supported by linear programming and cloud computing

    Get PDF
    With the increasing number of mobile subscribers worldwide, there is a need for fast and reliable algorithms for planning/optimization of mobile networks, especially because, in order to maintain a network’s quality of service, an operator might need to deploy more equipment. This paper presents a quick and reliable way to automatically plan a set of frequencies in a cellular network, using both cloud technologies and linear programming. We evaluate our pattern in a realistic scenario of a Global System for Mobile communications protocol (GSM) network and compare the results to another already implemented commercial tool. Results show that even though network quality was similar, our algorithm was twelve times faster and used four times less memory. It was also able to frequency plan seventy cells simultaneously in less than three minutes. This mechanism was successfully integrated in the professional tool Metric, and is currently being used for cellular planning. Its extension for application to 3/4/5G networks is under study.info:eu-repo/semantics/publishedVersio

    Sublinear Computation Paradigm

    Get PDF
    This open access book gives an overview of cutting-edge work on a new paradigm called the “sublinear computation paradigm,” which was proposed in the large multiyear academic research project “Foundations of Innovative Algorithms for Big Data.” That project ran from October 2014 to March 2020, in Japan. To handle the unprecedented explosion of big data sets in research, industry, and other areas of society, there is an urgent need to develop novel methods and approaches for big data analysis. To meet this need, innovative changes in algorithm theory for big data are being pursued. For example, polynomial-time algorithms have thus far been regarded as “fast,” but if a quadratic-time algorithm is applied to a petabyte-scale or larger big data set, problems are encountered in terms of computational resources or running time. To deal with this critical computational and algorithmic bottleneck, linear, sublinear, and constant time algorithms are required. The sublinear computation paradigm is proposed here in order to support innovation in the big data era. A foundation of innovative algorithms has been created by developing computational procedures, data structures, and modelling techniques for big data. The project is organized into three teams that focus on sublinear algorithms, sublinear data structures, and sublinear modelling. The work has provided high-level academic research results of strong computational and algorithmic interest, which are presented in this book. The book consists of five parts: Part I, which consists of a single chapter on the concept of the sublinear computation paradigm; Parts II, III, and IV review results on sublinear algorithms, sublinear data structures, and sublinear modelling, respectively; Part V presents application results. The information presented here will inspire the researchers who work in the field of modern algorithms

    27th Annual European Symposium on Algorithms: ESA 2019, September 9-11, 2019, Munich/Garching, Germany

    Get PDF
    corecore