30 research outputs found
First-order Convex Optimization Methods for Signal and Image Processing
In this thesis we investigate the use of first-order convex optimization methods applied to problems in signal and image processing. First we make a general introduction to convex optimization, first-order methods and their iteration com-plexity. Then we look at different techniques, which can be used with first-order methods such as smoothing, Lagrange multipliers and proximal gradient meth-ods. We continue by presenting different applications of convex optimization and notable convex formulations with an emphasis on inverse problems and sparse signal processing. We also describe the multiple-description problem. We finally present the contributions of the thesis. The remaining parts of the thesis consist of five research papers. The first paper addresses non-smooth first-order convex optimization and the trade-off between accuracy and smoothness of the approximating smooth function. The second and third papers concern discrete linear inverse problems and reliable numerical reconstruction software. The last two papers present a convex opti-mization formulation of the multiple-description problem and a method to solve it in the case of large-scale instances. i i
Parameter Selection and Pre-Conditioning for a Graph Form Solver
In a recent paper, Parikh and Boyd describe a method for solving a convex
optimization problem, where each iteration involves evaluating a proximal
operator and projection onto a subspace. In this paper we address the critical
practical issues of how to select the proximal parameter in each iteration, and
how to scale the original problem variables, so as the achieve reliable
practical performance. The resulting method has been implemented as an
open-source software package called POGS (Proximal Graph Solver), that targets
multi-core and GPU-based systems, and has been tested on a wide variety of
practical problems. Numerical results show that POGS can solve very large
problems (with, say, more than a billion coefficients in the data), to modest
accuracy in a few tens of seconds. As just one example, a radiation treatment
planning problem with around 100 million coefficients in the data can be solved
in a few seconds, as compared to around one hour with an interior-point method.Comment: 28 pages, 1 figure, 1 open source implementatio
Multilevel optimisation for computer vision
The recent spark in machine learning and computer vision methods requiring increasingly larger datasets has motivated the introduction of optimisation algorithms specifically tailored to solve very large problems within practical time constraints. This demand in algorithms challenges the practicability of state of the art methods requiring new approaches that can take advantage of not only the problem’s mathematical structure, but also its data structure. Fortunately, such structure is present in many computer vision applications, where the problems can be modelled with varying degrees of fidelity. This structure suggests using multiscale models and thus multilevel algorithms.
The objective of this thesis is to develop, implement and test provably convergent multilevel optimisation algorithms for convex composite optimisation problems in general and its applications in computer vision in particular. Our first multilevel algorithm solves convex composite optimisation problem and it is most efficient particularly for the robust facial recognition task. The method uses concepts from proximal gradient, mirror descent and multilevel optimisation algorithms, thus we call it multilevel accelerated gradient mirror descent algorithm (MAGMA). We first show that MAGMA has the same theoretical convergence rate as the state of the art first order methods and has much lower per iteration complexity. Then we demonstrate its practical advantage on many facial recognition problems. The second part of the thesis introduces new multilevel procedure most appropriate for the robust PCA problems requiring iterative SVD computations. We propose to exploit the multiscale structure present in these problems by constructing lower dimensional matrices and use its singular values for each iteration of the optimisation procedure. We implement this approach on three different optimisation algorithms - inexact ALM, Frank-Wolfe Thresholding and non-convex alternating projections. In this case as well we show that these multilevel algorithms converge (to an exact or approximate) solution with the same convergence rate as their standard counterparts and test all three methods on numerous synthetic and real life problems demonstrating that the multilevel algorithms are not only much faster, but also solve problems that often cannot be solved by their standard counterparts.Open Acces
A convergent blind deconvolution method for post-adaptive-optics astronomical imaging
In this paper we propose a blind deconvolution method which applies to data
perturbed by Poisson noise. The objective function is a generalized
Kullback-Leibler divergence, depending on both the unknown object and unknown
point spread function (PSF), without the addition of regularization terms;
constrained minimization, with suitable convex constraints on both unknowns, is
considered. The problem is nonconvex and we propose to solve it by means of an
inexact alternating minimization method, whose global convergence to stationary
points of the objective function has been recently proved in a general setting.
The method is iterative and each iteration, also called outer iteration,
consists of alternating an update of the object and the PSF by means of fixed
numbers of iterations, also called inner iterations, of the scaled gradient
projection (SGP) method. The use of SGP has two advantages: first, it allows to
prove global convergence of the blind method; secondly, it allows the
introduction of different constraints on the object and the PSF. The specific
constraint on the PSF, besides non-negativity and normalization, is an upper
bound derived from the so-called Strehl ratio, which is the ratio between the
peak value of an aberrated versus a perfect wavefront. Therefore a typical
application is the imaging of modern telescopes equipped with adaptive optics
systems for partial correction of the aberrations due to atmospheric
turbulence. In the paper we describe the algorithm and we recall the results
leading to its convergence. Moreover we illustrate its effectiveness by means
of numerical experiments whose results indicate that the method, pushed to
convergence, is very promising in the reconstruction of non-dense stellar
clusters. The case of more complex astronomical targets is also considered, but
in this case regularization by early stopping of the outer iterations is
required
A Riemannian Proximal Newton Method
In recent years, the proximal gradient method and its variants have been
generalized to Riemannian manifolds for solving optimization problems with an
additively separable structure, i.e., , where is continuously
differentiable, and may be nonsmooth but convex with computationally
reasonable proximal mapping. In this paper, we generalize the proximal Newton
method to embedded submanifolds for solving the type of problem with . The generalization relies on the Weingarten and semismooth
analysis. It is shown that the Riemannian proximal Newton method has a local
superlinear convergence rate under certain reasonable assumptions. Moreover, a
hybrid version is given by concatenating a Riemannian proximal gradient method
and the Riemannian proximal Newton method. It is shown that if the objective
function satisfies the Riemannian KL property and the switch parameter is
chosen appropriately, then the hybrid method converges globally and also has a
local superlinear convergence rate. Numerical experiments on random and
synthetic data are used to demonstrate the performance of the proposed methods
Acceleration Methods for MRI
Acceleration methods are a critical area of research for MRI. Two of the most important acceleration techniques involve parallel imaging and compressed sensing. These advanced signal processing techniques have the potential to drastically reduce scan times and provide radiologists with new information for diagnosing disease. However, many of these new techniques require solving difficult optimization problems, which motivates the development of more advanced algorithms to solve them. In addition, acceleration methods have not reached maturity in some applications, which motivates the development of new models tailored to these applications. This dissertation makes advances in three different areas of accelerations. The first is the development of a new algorithm (called B1-Based, Adaptive Restart, Iterative Soft Thresholding Algorithm or BARISTA), that solves a parallel MRI optimization problem with compressed sensing assumptions. BARISTA is shown to be 2-3 times faster and more robust to parameter selection than current state-of-the-art variable splitting methods. The second contribution is the extension of BARISTA ideas to non-Cartesian trajectories that also leads to a 2-3 times acceleration over previous methods. The third contribution is the development of a new model for functional MRI that enables a 3-4 factor of acceleration of effective temporal resolution in functional MRI scans. Several variations of the new model are proposed, with an ROC curve analysis showing that a combination low-rank/sparsity model giving the best performance in identifying the resting-state motor network.PhDBiomedical EngineeringUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttp://deepblue.lib.umich.edu/bitstream/2027.42/120841/1/mmuckley_1.pd