219 research outputs found
Mirror Descent and Convex Optimization Problems With Non-Smooth Inequality Constraints
We consider the problem of minimization of a convex function on a simple set
with convex non-smooth inequality constraint and describe first-order methods
to solve such problems in different situations: smooth or non-smooth objective
function; convex or strongly convex objective and constraint; deterministic or
randomized information about the objective and constraint. We hope that it is
convenient for a reader to have all the methods for different settings in one
place. Described methods are based on Mirror Descent algorithm and switching
subgradient scheme. One of our focus is to propose, for the listed different
settings, a Mirror Descent with adaptive stepsizes and adaptive stopping rule.
This means that neither stepsize nor stopping rule require to know the
Lipschitz constant of the objective or constraint. We also construct Mirror
Descent for problems with objective function, which is not Lipschitz
continuous, e.g. is a quadratic function. Besides that, we address the problem
of recovering the solution of the dual problem
Plasma formation from ultracold Rydberg gases
Recent experiments have demonstrated the spontaneous evolution of a gas of
ultracold Rydberg atoms into an expanding ultracold plasma, as well as the
reverse process of plasma recombination into highly excited atomic states.
Treating the evolution of the plasma on the basis of kinetic equations, while
ionization/excitation and recombination are incorporated using rate equations,
we have investigated theoretically the Rydberg-to-plasma transition. Including
the influence of spatial correlations on the plasma dynamics in an approximate
way we find that ionic correlations change the results only quantitatively but
not qualitatively
A new picture of the Lifshitz critical behavior
New field theoretic renormalization group methods are developed to describe
in a unified fashion the critical exponents of an m-fold Lifshitz point at the
two-loop order in the anisotropic (m not equal to d) and isotropic (m=d close
to 8) situations. The general theory is illustrated for the N-vector phi^4
model describing a d-dimensional system. A new regularization and
renormalization procedure is presented for both types of Lifshitz behavior. The
anisotropic cases are formulated with two independent renormalization group
transformations. The description of the isotropic behavior requires only one
type of renormalization group transformation. We point out the conceptual
advantages implicit in this picture and show how this framework is related to
other previous renormalization group treatments for the Lifshitz problem. The
Feynman diagrams of arbitrary loop-order can be performed analytically provided
these integrals are considered to be homogeneous functions of the external
momenta scales. The anisotropic universality class (N,d,m) reduces easily to
the Ising-like (N,d) when m=0. We show that the isotropic universality class
(N,m) when m is close to 8 cannot be obtained from the anisotropic one in the
limit d --> m near 8. The exponents for the uniaxial case d=3, N=m=1 are in
good agreement with recent Monte Carlo simulations for the ANNNI model.Comment: 48 pages, no figures, two typos fixe
Implementation of an Optimal First-Order Method for Strongly Convex Total Variation Regularization
We present a practical implementation of an optimal first-order method, due
to Nesterov, for large-scale total variation regularization in tomographic
reconstruction, image deblurring, etc. The algorithm applies to -strongly
convex objective functions with -Lipschitz continuous gradient. In the
framework of Nesterov both and are assumed known -- an assumption
that is seldom satisfied in practice. We propose to incorporate mechanisms to
estimate locally sufficient and during the iterations. The mechanisms
also allow for the application to non-strongly convex functions. We discuss the
iteration complexity of several first-order methods, including the proposed
algorithm, and we use a 3D tomography problem to compare the performance of
these methods. The results show that for ill-conditioned problems solved to
high accuracy, the proposed method significantly outperforms state-of-the-art
first-order methods, as also suggested by theoretical results.Comment: 23 pages, 4 figure
Recommended from our members
PKCθ Regulates T-Cell Leukemia-Initiating Activity via Reactive Oxygen Species
Reactive oxygen species (ROS), a by-product of cellular metabolism, damage intracellular macromolecules and, in excess, can promote normal hematopoietic stem cell differentiation and exhaustion1–3. However, mechanisms that regulate ROS levels in leukemia-initiating cells (LICs) and the biological role of ROS in these cells remain largely unknown. We show here the ROSlow subset of CD44+ cells in T-cell acute lymphoblastic leukemia (T-ALL), a malignancy of immature T-cell progenitors, to be highly enriched in the most aggressive LICs, and that ROS are maintained at low levels by downregulation of protein kinase C theta (PKCθ). Strikingly, primary mouse T-ALLs lacking PKCθ show improved LIC activity whereas enforced PKCθ expression in both mouse and human primary T-ALLs compromised LIC activity. We also demonstrate that PKCθ is positively regulated by RUNX1, and that NOTCH1, which is frequently activated by mutation in T-ALL4–6 and required for LIC activity in both mouse and human models7,8, downregulates PKCθ and ROS via a novel pathway involving induction of RUNX3 and subsequent repression of RUNX1. These results reveal key functional roles for PKCθ and ROS in T-ALL and suggest that aggressive biological behavior in vivo could be limited by therapeutic strategies that promote PKCθ expression/activity or ROS accumulation
Advances in low-memory subgradient optimization
One of the main goals in the development of non-smooth optimization is to cope with high dimensional problems by decomposition, duality or Lagrangian relaxation which greatly reduces the number of variables at the cost of worsening differentiability of objective or constraints. Small or medium dimensionality of resulting non-smooth problems allows to use bundle-type algorithms to achieve higher rates of convergence and obtain higher accuracy, which of course came at the cost of additional memory requirements, typically of the order of n2, where n is the number of variables of non-smooth problem. However with the rapid development of more and more sophisticated models in industry, economy, finance, et all such memory requirements are becoming too hard to satisfy. It raised the interest in subgradient-based low-memory algorithms and later developments in this area significantly improved over their early variants still preserving O(n) memory requirements. To review these developments this chapter is devoted to the black-box subgradient algorithms with the minimal requirements for the storage of auxiliary results, which are necessary to execute these algorithms. To provide historical perspective this survey starts with the original result of N.Z. Shor which opened this field with the application to the classical transportation problem. The theoretical complexity bounds for smooth and non-smooth convex and quasi-convex optimization problems are briefly exposed in what follows to introduce to the relevant fundamentals of non-smooth optimization. Special attention in this section is given to the adaptive step-size policy which aims to attain lowest complexity bounds. Unfortunately the non-differentiability of objective function in convex optimization essentially slows down the theoretical low bounds for the rate of convergence in subgradient optimization compared to the smooth case but there are different modern techniques that allow to solve non-smooth convex optimization problems faster then dictate lower complexity bounds. In this work the particular attention is given to Nesterov smoothing technique, Nesterov Universal approach, and Legendre (saddle point) representation approach. The new results on Universal Mirror Prox algorithms represent the original parts of the survey. To demonstrate application of non-smooth convex optimization algorithms for solution of huge-scale extremal problems we consider convex optimization problems with non-smooth functional constraints and propose two adaptive Mirror Descent methods. The first method is of primal-dual variety and proved to be optimal in terms of lower oracle bounds for the class of Lipschitz-continuous convex objective and constraints. The advantages of application of this method to sparse Truss Topology Design problem are discussed in certain details. The second method can be applied for solution of convex and quasi-convex optimization problems and is optimal in a sense of complexity bounds. The conclusion part of the survey contains the important references that characterize recent developments of non-smooth convex optimization
The Use of Polls to Enhance Formative Assessment Processes in Mathematics Classroom Discussions
This contribution addresses the theme of technology for formative assessment in the mathematics classroom and in particular the ways connected classroom technology may support formative assessment strategies in whole class activities. Design experiments have been developed through the use of a connected classroom technology by which students may share their productions, opinions, and reflections with their classmates and the teacher during or at the end of a mathematical activity. With this technology the teacher may create polls, submit them to the students, gather their answers and show the results in real time. The paper discusses how polls can be used during classroom activities to foster the activation of formative assessment strategies. As a result of the design-based research, a classification of polls according to their contents and aims is proposed. Different ways of structuring classroom discussions and patterns of formative assessment strategies, which are developed from the different types of polls, are discussed
Towards Machine Wald
The past century has seen a steady increase in the need of estimating and
predicting complex systems and making (possibly critical) decisions with
limited information. Although computers have made possible the numerical
evaluation of sophisticated statistical models, these models are still designed
\emph{by humans} because there is currently no known recipe or algorithm for
dividing the design of a statistical model into a sequence of arithmetic
operations. Indeed enabling computers to \emph{think} as \emph{humans} have the
ability to do when faced with uncertainty is challenging in several major ways:
(1) Finding optimal statistical models remains to be formulated as a well posed
problem when information on the system of interest is incomplete and comes in
the form of a complex combination of sample data, partial knowledge of
constitutive relations and a limited description of the distribution of input
random variables. (2) The space of admissible scenarios along with the space of
relevant information, assumptions, and/or beliefs, tend to be infinite
dimensional, whereas calculus on a computer is necessarily discrete and finite.
With this purpose, this paper explores the foundations of a rigorous framework
for the scientific computation of optimal statistical estimators/models and
reviews their connections with Decision Theory, Machine Learning, Bayesian
Inference, Stochastic Optimization, Robust Optimization, Optimal Uncertainty
Quantification and Information Based Complexity.Comment: 37 page
- …