3,148 research outputs found
Conditional gradient type methods for composite nonlinear and stochastic optimization
In this paper, we present a conditional gradient type (CGT) method for
solving a class of composite optimization problems where the objective function
consists of a (weakly) smooth term and a (strongly) convex regularization term.
While including a strongly convex term in the subproblems of the classical
conditional gradient (CG) method improves its rate of convergence, it does not
cost per iteration as much as general proximal type algorithms. More
specifically, we present a unified analysis for the CGT method in the sense
that it achieves the best-known rate of convergence when the weakly smooth term
is nonconvex and possesses (nearly) optimal complexity if it turns out to be
convex. While implementation of the CGT method requires explicitly estimating
problem parameters like the level of smoothness of the first term in the
objective function, we also present a few variants of this method which relax
such estimation. Unlike general proximal type parameter free methods, these
variants of the CGT method do not require any additional effort for computing
(sub)gradients of the objective function and/or solving extra subproblems at
each iteration. We then generalize these methods under stochastic setting and
present a few new complexity results. To the best of our knowledge, this is the
first time that such complexity results are presented for solving stochastic
weakly smooth nonconvex and (strongly) convex optimization problems
Zeroth-order Nonconvex Stochastic Optimization: Handling Constraints, High-Dimensionality and Saddle-Points
In this paper, we propose and analyze zeroth-order stochastic approximation
algorithms for nonconvex and convex optimization, with a focus on addressing
constrained optimization, high-dimensional setting and saddle-point avoiding.
To handle constrained optimization, we first propose generalizations of the
conditional gradient algorithm achieving rates similar to the standard
stochastic gradient algorithm using only zeroth-order information. To
facilitate zeroth-order optimization in high-dimensions, we explore the
advantages of structural sparsity assumptions. Specifically, (i) we highlight
an implicit regularization phenomenon where the standard stochastic gradient
algorithm with zeroth-order information adapts to the sparsity of the problem
at hand by just varying the step-size and (ii) propose a truncated stochastic
gradient algorithm with zeroth-order information, whose rate of convergence
depends only poly-logarithmically on the dimensionality. We next focus on
avoiding saddle-points in non-convex setting. Towards that, we interpret the
Gaussian smoothing technique for estimating gradient based on zeroth-order
information as an instantiation of first-order Stein's identity. Based on this,
we provide a novel linear-(in dimension) time estimator of the Hessian matrix
of a function using only zeroth-order information, which is based on
second-order Stein's identity. We then provide an algorithm for avoiding
saddle-points, which is based on a zeroth-order cubic regularization Newton's
method and discuss its convergence rates
Accelerated Gradient Methods for Nonconvex Nonlinear and Stochastic Programming
In this paper, we generalize the well-known Nesterov's accelerated gradient
(AG) method, originally designed for convex smooth optimization, to solve
nonconvex and possibly stochastic optimization problems. We demonstrate that by
properly specifying the stepsize policy, the AG method exhibits the best known
rate of convergence for solving general nonconvex smooth optimization problems
by using first-order information, similarly to the gradient descent method. We
then consider an important class of composite optimization problems and show
that the AG method can solve them uniformly, i.e., by using the same aggressive
stepsize policy as in the convex case, even if the problem turns out to be
nonconvex. We demonstrate that the AG method exhibits an optimal rate of
convergence if the composite problem is convex, and improves the best known
rate of convergence if the problem is nonconvex. Based on the AG method, we
also present new nonconvex stochastic approximation methods and show that they
can improve a few existing rates of convergence for nonconvex stochastic
optimization. To the best of our knowledge, this is the first time that the
convergence of the AG method has been established for solving nonconvex
nonlinear programming in the literature
Majorana Zero-Energy Mode and Fractal Structure in Fibonacci-Kitaev Chain
We theoretically study a Kitaev chain with a quasiperiodic potential, where
the quasiperiodicity is introduced by a Fibonacci sequence. Based on an
analysis of the Majorana zero-energy mode, we find the critical -wave
superconducting pairing potential separating a topological phase and a
non-topological phase. The topological phase diagram with respect to Fibonacci
potentials follow a self-similar fractal structure characterized by the
box-counting dimension, which is an example of the interplay of fractal and
topology like the Hofstadter's butterfly in quantum Hall insulators.Comment: 5 pages, 4 figure
Mini-batch Stochastic Approximation Methods for Nonconvex Stochastic Composite Optimization
This paper considers a class of constrained stochastic composite optimization
problems whose objective function is given by the summation of a differentiable
(possibly nonconvex) component, together with a certain non-differentiable (but
convex) component. In order to solve these problems, we propose a randomized
stochastic projected gradient (RSPG) algorithm, in which proper mini-batch of
samples are taken at each iteration depending on the total budget of stochastic
samples allowed. The RSPG algorithm also employs a general distance function to
allow taking advantage of the geometry of the feasible region. Complexity of
this algorithm is established in a unified setting, which shows nearly optimal
complexity of the algorithm for convex stochastic programming. A
post-optimization phase is also proposed to significantly reduce the variance
of the solutions returned by the algorithm. In addition, based on the RSPG
algorithm, a stochastic gradient free algorithm, which only uses the stochastic
zeroth-order information, has been also discussed. Some preliminary numerical
results are also provided.Comment: 32 page
Accelerated Gradient Methods for Networked Optimization
We develop multi-step gradient methods for network-constrained optimization
of strongly convex functions with Lipschitz-continuous gradients. Given the
topology of the underlying network and bounds on the Hessian of the objective
function, we determine the algorithm parameters that guarantee the fastest
convergence and characterize situations when significant speed-ups can be
obtained over the standard gradient method. Furthermore, we quantify how the
performance of the gradient method and its accelerated counterpart are affected
by uncertainty in the problem data, and conclude that in most cases our
proposed method outperforms gradient descent. Finally, we apply the proposed
technique to three engineering problems: resource allocation under network-wide
budget constraints, distributed averaging, and Internet congestion control. In
all cases, we demonstrate that our algorithm converges more rapidly than
alternative algorithms reported in the literature
Non-asymptotic Results for Langevin Monte Carlo: Coordinate-wise and Black-box Sampling
Discretization of continuous-time diffusion processes, using gradient and
Hessian information, is a popular technique for sampling. For example, the
Euler-Maruyama discretization of the Langevin diffusion process, called as
Langevin Monte Carlo (LMC), is a canonical algorithm for sampling from strongly
log-concave densities. In this work, we make several theoretical contributions
to the literature on such sampling techniques. Specifically, we first provide a
Randomized Coordinate-wise LMC algorithm suitable for large-scale sampling
problems and provide a theoretical analysis. We next consider the case of
zeroth-order or black-box sampling where one only obtains evaluates of the
density. Based on Gaussian Stein's identities we then estimate the gradient and
Hessian information and leverage it in the context of black-box sampling. We
then provide a theoretical analysis of gradient and Hessian based
discretizations of Langevin and kinetic Langevin diffusion processes for
sampling, quantifying the non-asymptotic accuracy. We also consider
high-dimensional black-box sampling under the assumption that the density
depends only on a small subset of the entire coordinates. We propose a variable
selection technique based on zeroth-order gradient estimates and establish its
theoretical guarantees. Our theoretical contributions extend the practical
applicability of sampling algorithms to the large-scale, black-box and
high-dimensional settings
Generalized Uniformly Optimal Methods for Nonlinear Programming
In this paper, we present a generic framework to extend existing uniformly
optimal convex programming algorithms to solve more general nonlinear, possibly
nonconvex, optimization problems. The basic idea is to incorporate a local
search step (gradient descent or Quasi-Newton iteration) into these uniformly
optimal convex programming methods, and then enforce a monotone decreasing
property of the function values computed along the trajectory. Algorithms of
these types will then achieve the best known complexity for nonconvex problems,
and the optimal complexity for convex ones without requiring any problem
parameters. As a consequence, we can have a unified treatment for a general
class of nonlinear programming problems regardless of their convexity and
smoothness level. In particular, we show that the accelerated gradient and
level methods, both originally designed for solving convex optimization
problems only, can be used for solving both convex and nonconvex problems
uniformly. In a similar vein, we show that some well-studied techniques for
nonlinear programming, e.g., Quasi-Newton iteration, can be embedded into
optimal convex optimization algorithms to possibly further enhance their
numerical performance. Our theoretical and algorithmic developments are
complemented by some promising numerical results obtained for solving a few
important nonconvex and nonlinear data analysis problems in the literature
Mean-field study of the Bose-Hubbard model in Penrose lattice
We examine the Bose-Hubbard model in the Penrose lattice based on
inhomogeneous mean-field theory. Since averaged coordination number in the
Penrose lattice is four, mean-field phase diagram consisting of the Mott
insulator (MI) and superfluid (SF) phase is similar to that of the square
lattice. However, the spatial distribution of Bose condensate in the SF phase
is significantly different from uniform distribution in the square lattice. We
find a fractal structure in its distribution near the MI-SF phase boundary. The
emergence of the fractal structure is a consequence of cooperative effect
between quasiperiodicity in the Penrose lattice and criticality at the phase
transition.Comment: 9 pages, 5 figures, 1 tabl
A more robust multiparameter conformal mapping method for geometry generation of any arbitrary ship section
The central problem of strip theory is the calculation of potential
flowaround 2D sections. One particular method of solutions to this problem is
conformal mapping of the body section to the unit circle over which a solution
of potential flow is available. Here, a new multiparameter conformal mapping
method is presented that can map any arbitrary section onto a unit circle with
good accuracy. The procedure for finding the corresponding mapping coefficients
is iterative. The suggested mapping technique is shown to be capable of
appropriately mapping any chined, bulbous, and large and fine sections. Several
examples of mapping symmetric and nonsymmetric sections are demonstrated. For
symmetric and nonsymmetric sections, the results of the current method are
compared against other mapping techniques, and the currently produced
geometries display good agreement with the actual geometries.Comment: 40 pages, 31 figures, 4 Table
- …