110 research outputs found

    An Expanded Theoretical Treatment of Iteration-Dependent Majorize-Minimize Algorithms

    Full text link
    The majorize-minimize (MM) optimization technique has received considerable attention in signal and image processing applications, as well as in statistics literature. At each iteration of an MM algorithm, one constructs a tangent majorant function that majorizes the given cost function and is equal to it at the current iterate. The next iterate is obtained by minimizing this tangent majorant function, resulting in a sequence of iterates that reduces the cost function monotonically. A well-known special case of MM methods are expectation-maximization algorithms. In this paper, we expand on previous analyses of MM, due to Fessler and Hero, that allowed the tangent majorants to be constructed in iteration-dependent ways. Also, this paper overcomes an error in one of those earlier analyses. There are three main aspects in which our analysis builds upon previous work. First, our treatment relaxes many assumptions related to the structure of the cost function, feasible set, and tangent majorants. For example, the cost function can be nonconvex and the feasible set for the problem can be any convex set. Second, we propose convergence conditions, based on upper curvature bounds, that can be easier to verify than more standard continuity conditions. Furthermore, these conditions allow for considerable design freedom in the iteration-dependent behavior of the algorithm. Finally, we give an original characterization of the local region of convergence of MM algorithms based on connected (e.g., convex) tangent majorants. For such algorithms, cost function minimizers will locally attract the iterates over larger neighborhoods than typically is guaranteed with other methods. This expanded treatment widens the scope of the MM algorithm designs that can be considered for signal and image processing applications, allows us to verify the convergent behavior of previously published algorithms, and gives a fuller understanding overall of how these algorithms behave.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/85958/1/Fessler34.pd

    ๋ณ‘๋ ฌํ™” ์šฉ์ดํ•œ ํ†ต๊ณ„๊ณ„์‚ฐ ๋ฐฉ๋ฒ•๋ก ๊ณผ ํ˜„๋Œ€ ๊ณ ์„ฑ๋Šฅ ์ปดํ“จํŒ… ํ™˜๊ฒฝ์—์˜ ์ ์šฉ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ์ž์—ฐ๊ณผํ•™๋Œ€ํ•™ ํ†ต๊ณ„ํ•™๊ณผ, 2020. 8. ์›์ค‘ํ˜ธ.Technological advances in the past decade, hardware and software alike, have made access to high-performance computing (HPC) easier than ever. In this dissertation, easily-parallelizable, inversion-free, and variable-separated algorithms and their implementation in statistical computing are discussed. The first part considers statistical estimation problems under structured sparsity posed as minimization of a sum of two or three convex functions, one of which is a composition of non-smooth and linear functions. Examples include graph-guided sparse fused lasso and overlapping group lasso. Two classes of inversion-free primal-dual algorithms are considered and unified from a perspective of monotone operator theory. From this unification, a continuum of preconditioned forward-backward operator splitting algorithms amenable to parallel and distributed computing is proposed. The unification is further exploited to introduce a continuum of accelerated algorithms on which the theoretically optimal asymptotic rate of convergence is obtained. For the second part, easy-to-use distributed matrix data structures in PyTorch and Julia are presented. They enable users to write code once and run it anywhere from a laptop to a workstation with multiple graphics processing units (GPUs) or a supercomputer in a cloud. With these data structures, various parallelizable statistical applications, including nonnegative matrix factorization, positron emission tomography, multidimensional scaling, and โ„“1-regularized Cox regression, are demonstrated. The examples scale up to an 8-GPU workstation and a 720-CPU-core cluster in a cloud. As a case in point, the onset of type-2 diabetes from the UK Biobank with 400,000 subjects and about 500,000 single nucleotide polymorphisms is analyzed using the HPC โ„“1-regularized Cox regression. Fitting a half-million variate model took about 50 minutes, reconfirming known associations. To my knowledge, the feasibility of a joint genome-wide association analysis of survival outcomes at this scale is first demonstrated.์ง€๋‚œ 10๋…„๊ฐ„์˜ ํ•˜๋“œ์›จ์–ด์™€ ์†Œํ”„ํŠธ์›จ์–ด์˜ ๊ธฐ์ˆ ์ ์ธ ๋ฐœ์ „์€ ๊ณ ์„ฑ๋Šฅ ์ปดํ“จํŒ…์˜ ์ ‘๊ทผ์žฅ๋ฒฝ์„ ๊ทธ ์–ด๋Š ๋•Œ๋ณด๋‹ค ๋‚ฎ์ถ”์—ˆ๋‹ค. ์ด ํ•™์œ„๋…ผ๋ฌธ์—์„œ๋Š” ๋ณ‘๋ ฌํ™” ์šฉ์ดํ•˜๊ณ  ์—ญํ–‰๋ ฌ ์—ฐ์‚ฐ์ด ์—†๋Š” ๋ณ€์ˆ˜ ๋ถ„๋ฆฌ ์•Œ๊ณ ๋ฆฌ์ฆ˜๊ณผ ๊ทธ ํ†ต๊ณ„๊ณ„์‚ฐ์—์„œ์˜ ๊ตฌํ˜„์„ ๋…ผ์˜ํ•œ๋‹ค. ์ฒซ ๋ถ€๋ถ„์€ ๋ณผ๋ก ํ•จ์ˆ˜ ๋‘ ๊ฐœ ๋˜๋Š” ์„ธ ๊ฐœ์˜ ํ•ฉ์œผ๋กœ ๋‚˜ํƒ€๋‚˜๋Š” ๊ตฌ์กฐํ™”๋œ ํฌ์†Œ ํ†ต๊ณ„ ์ถ”์ • ๋ฌธ์ œ์— ๋Œ€ํ•ด ๋‹ค๋ฃฌ๋‹ค. ์ด ๋•Œ ํ•จ์ˆ˜๋“ค ์ค‘ ํ•˜๋‚˜๋Š” ๋น„ํ‰ํ™œ ํ•จ์ˆ˜์™€ ์„ ํ˜• ํ•จ์ˆ˜์˜ ํ•ฉ์„ฑ์œผ๋กœ ๋‚˜ํƒ€๋‚œ๋‹ค. ๊ทธ ์˜ˆ์‹œ๋กœ๋Š” ๊ทธ๋ž˜ํ”„ ๊ตฌ์กฐ๋ฅผ ํ†ตํ•ด ์œ ๋„๋˜๋Š” ํฌ์†Œ ์œตํ•ฉ Lasso ๋ฌธ์ œ์™€ ํ•œ ๋ณ€์ˆ˜๊ฐ€ ์—ฌ๋Ÿฌ ๊ทธ๋ฃน์— ์†ํ•  ์ˆ˜ ์žˆ๋Š” ๊ทธ๋ฃน Lasso ๋ฌธ์ œ๊ฐ€ ์žˆ๋‹ค. ์ด๋ฅผ ํ’€๊ธฐ ์œ„ํ•ด ์—ญํ–‰๋ ฌ ์—ฐ์‚ฐ์ด ์—†๋Š” ๋‘ ์ข…๋ฅ˜์˜ ์›์‹œ-์Œ๋Œ€ (primal-dual) ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ๋‹จ์กฐ ์—ฐ์‚ฐ์ž ์ด๋ก  ๊ด€์ ์—์„œ ํ†ตํ•ฉํ•˜๋ฉฐ ์ด๋ฅผ ํ†ตํ•ด ๋ณ‘๋ ฌํ™” ์šฉ์ดํ•œ precondition๋œ ์ „๋ฐฉ-ํ›„๋ฐฉ ์—ฐ์‚ฐ์ž ๋ถ„ํ•  ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ ์ง‘ํ•ฉ์„ ์ œ์•ˆํ•œ๋‹ค. ์ด ํ†ตํ•ฉ์€ ์ ๊ทผ์ ์œผ๋กœ ์ตœ์  ์ˆ˜๋ ด๋ฅ ์„ ๊ฐ–๋Š” ๊ฐ€์† ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ ์ง‘ํ•ฉ์„ ๊ตฌ์„ฑํ•˜๋Š” ๋ฐ ํ™œ์šฉ๋œ๋‹ค. ๋‘ ๋ฒˆ์งธ ๋ถ€๋ถ„์—์„œ๋Š” PyTorch์™€ Julia๋ฅผ ํ†ตํ•ด ์‚ฌ์šฉํ•˜๊ธฐ ์‰ฌ์šด ๋ถ„์‚ฐ ํ–‰๋ ฌ ์ž๋ฃŒ ๊ตฌ์กฐ๋ฅผ ์ œ์‹œํ•œ๋‹ค. ์ด ๊ตฌ์กฐ๋Š” ์‚ฌ์šฉ์ž๋“ค์ด ์ฝ”๋“œ๋ฅผ ํ•œ ๋ฒˆ ์ž‘์„ฑํ•˜๋ฉด ์ด๊ฒƒ์„ ๋…ธํŠธ๋ถ ํ•œ ๋Œ€์—์„œ๋ถ€ํ„ฐ ์—ฌ๋Ÿฌ ๋Œ€์˜ ๊ทธ๋ž˜ํ”ฝ ์ฒ˜๋ฆฌ ์žฅ์น˜ (GPU)๋ฅผ ๊ฐ€์ง„ ์›Œํฌ์Šคํ…Œ์ด์…˜, ๋˜๋Š” ํด๋ผ์šฐ๋“œ ์ƒ์— ์žˆ๋Š” ์Šˆํผ์ปดํ“จํ„ฐ๊นŒ์ง€ ๋‹ค์–‘ํ•œ ์Šค์ผ€์ผ์—์„œ ์‹คํ–‰ํ•  ์ˆ˜ ์žˆ๊ฒŒ ํ•ด ์ค€๋‹ค. ์•„์šธ๋Ÿฌ, ์ด ์ž๋ฃŒ ๊ตฌ์กฐ๋ฅผ ๋น„์Œ ํ–‰๋ ฌ ๋ถ„ํ•ด, ์–‘์ „์ž ๋‹จ์ธต ์ดฌ์˜, ๋‹ค์ฐจ์› ์ฒ™ ๋„๋ฒ•, โ„“1-๋ฒŒ์ ํ™” Cox ํšŒ๊ท€ ๋ถ„์„ ๋“ฑ ๋‹ค์–‘ํ•œ ๋ณ‘๋ ฌํ™” ๊ฐ€๋Šฅํ•œ ํ†ต๊ณ„์  ๋ฌธ์ œ์— ์ ์šฉํ•œ๋‹ค. ์ด ์˜ˆ์‹œ๋“ค์€ 8๋Œ€์˜ GPU๊ฐ€ ์žˆ๋Š” ์›Œํฌ์Šคํ…Œ์ด์…˜๊ณผ 720๊ฐœ์˜ ์ฝ”์–ด๊ฐ€ ์žˆ๋Š” ํด๋ผ์šฐ๋“œ ์ƒ์˜ ๊ฐ€์ƒ ํด๋Ÿฌ์Šคํ„ฐ์—์„œ ํ™•์žฅ ๊ฐ€๋Šฅํ–ˆ๋‹ค. ํ•œ ์‚ฌ๋ก€๋กœ 400,000๋ช…์˜ ๋Œ€์ƒ๊ณผ 500,000๊ฐœ์˜ ๋‹จ์ผ ์—ผ๊ธฐ ๋‹คํ˜•์„ฑ ์ •๋ณด๊ฐ€ ์žˆ๋Š” UK Biobank ์ž๋ฃŒ์—์„œ์˜ ์ œ2ํ˜• ๋‹น๋‡จ๋ณ‘ (T2D) ๋ฐœ๋ณ‘ ๋‚˜์ด๋ฅผ โ„“1-๋ฒŒ์ ํ™” Cox ํšŒ๊ท€ ๋ชจํ˜•์„ ํ†ตํ•ด ๋ถ„์„ํ–ˆ๋‹ค. 500,000๊ฐœ์˜ ๋ณ€์ˆ˜๊ฐ€ ์žˆ๋Š” ๋ชจํ˜•์„ ์ ํ•ฉ์‹œํ‚ค๋Š” ๋ฐ 50๋ถ„ ๊ฐ€๋Ÿ‰์˜ ์‹œ๊ฐ„์ด ๊ฑธ๋ ธ์œผ๋ฉฐ ์•Œ๋ ค์ง„ T2D ๊ด€๋ จ ๋‹คํ˜•์„ฑ๋“ค์„ ์žฌํ™•์ธํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ์ด๋Ÿฌํ•œ ๊ทœ๋ชจ์˜ ์ „์œ ์ „์ฒด ๊ฒฐํ•ฉ ์ƒ์กด ๋ถ„์„์€ ์ตœ์ดˆ๋กœ ์‹œ๋„๋œ ๊ฒƒ์ด๋‹ค.Chapter1Prologue 1 1.1 Introduction 1 1.2 Accessible High-Performance Computing Systems 4 1.2.1 Preliminaries 4 1.2.2 Multiple CPU nodes: clusters, supercomputers, and clouds 7 1.2.3 Multi-GPU node 9 1.3 Highly Parallelizable Algorithms 12 1.3.1 MM algorithms 12 1.3.2 Proximal gradient descent 14 1.3.3 Proximal distance algorithm 16 1.3.4 Primal-dual methods 17 Chapter 2 Easily Parallelizable and Distributable Class of Algorithms for Structured Sparsity, with Optimal Acceleration 20 2.1 Introduction 20 2.2 Unification of Algorithms LV and CV (g โ‰ก 0) 30 2.2.1 Relation between Algorithms LV and CV 30 2.2.2 Unified algorithm class 34 2.2.3 Convergence analysis 35 2.3 Optimal acceleration 39 2.3.1 Algorithms 40 2.3.2 Convergence analysis 41 2.4 Stochastic optimal acceleration 45 2.4.1 Algorithm 45 2.4.2 Convergence analysis 47 2.5 Numerical experiments 50 2.5.1 Model problems 50 2.5.2 Convergence behavior 52 2.5.3 Scalability 62 2.6 Discussion 63 Chapter 3 Towards Unified Programming for High-Performance Statistical Computing Environments 66 3.1 Introduction 66 3.2 Related Software 69 3.2.1 Message-passing interface and distributed array interfaces 69 3.2.2 Unified array interfaces for CPU and GPU 69 3.3 Easy-to-use Software Libraries for HPC 70 3.3.1 Deep learning libraries and HPC 70 3.3.2 Case study: PyTorch versus TensorFlow 73 3.3.3 A brief introduction to PyTorch 76 3.3.4 A brief introduction to Julia 80 3.3.5 Methods and multiple dispatch 80 3.3.6 Multidimensional arrays 82 3.3.7 Matrix multiplication 83 3.3.8 Dot syntax for vectorization 86 3.4 Distributed matrix data structure 87 3.4.1 Distributed matrices in PyTorch: distmat 87 3.4.2 Distributed arrays in Julia: MPIArray 90 3.5 Examples 98 3.5.1 Nonnegative matrix factorization 100 3.5.2 Positron emission tomography 109 3.5.3 Multidimensional scaling 113 3.5.4 L1-regularized Cox regression 117 3.5.5 Genome-wide survival analysis of the UK Biobank dataset 121 3.6 Discussion 126 Chapter 4 Conclusion 131 Appendix A Monotone Operator Theory 134 Appendix B Proofs for Chapter II 139 B.1 Preconditioned forward-backward splitting 139 B.2 Optimal acceleration 147 B.3 Optimal stochastic acceleration 158 Appendix C AWS EC2 and ParallelCluster 168 C.1 Overview 168 C.2 Glossary 169 C.3 Prerequisites 172 C.4 Installation 173 C.5 Configuration 173 C.6 Creating, accessing, and destroying the cluster 178 C.7 Installation of libraries 178 C.8 Running a job 179 C.9 Miscellaneous 180 Appendix D Code for memory-efficient L1-regularized Cox proportional hazards model 182 Appendix E Details of SNPs selected in L1-regularized Cox regression 184 Bibliography 188 ๊ตญ๋ฌธ์ดˆ๋ก 212Docto

    A general framework for nonlinear multigrid inversion

    Full text link

    Doctor of Philosophy

    Get PDF
    dissertationX-ray computed tomography (CT) is a widely popular medical imaging technique that allows for viewing of in vivo anatomy and physiology. In order to produce high-quality images and provide reliable treatment, CT imaging requires the precise knowledge of t

    Fast voxel line update for time-space image reconstruction

    Full text link

    Mini-Workshop: Innovative Trends in the Numerical Analysis and Simulation of Kinetic Equations

    Get PDF
    In multiscale modeling hierarchy, kinetic theory plays a vital role in connecting microscopic Newtonian mechanics and macroscopic continuum mechanics. As computing power grows, numerical simulation of kinetic equations has become possible and undergone rapid development over the past decade. Yet the unique challenges arising in these equations, such as highdimensionality, multiple scales, random inputs, positivity, entropy dissipation, etc., call for new advances of numerical methods. This mini-workshop brought together both senior and junior researchers working on various fastpaced growing numerical aspects of kinetic equations. The topics include, but were not limited to, uncertainty quantification, structure-preserving methods, phase transitions, asymptotic-preserving schemes, and fast methods for kinetic equations

    Computational optimization methods for large-scale inverse problems

    Get PDF
    • โ€ฆ
    corecore