168,441 research outputs found

    Quasilinearization Method and Summation of the WKB Series

    Full text link
    Solutions obtained by the quasilinearization method (QLM) are compared with the WKB solutions. Expansion of the pp-th QLM iterate in powers of \hbar reproduces the structure of the WKB series generating an infinite number of the WKB terms with the first 2p2^p terms reproduced exactly. The QLM quantization condition leads to exact energies for the P\"{o}schl-Teller, Hulthen, Hylleraas, Morse, Eckart potentials etc. For other, more complicated potentials the first QLM iterate, given by the closed analytic expression, is extremely accurate. The iterates converge very fast. The sixth iterate of the energy for the anharmonic oscillator and for the two-body Coulomb Dirac equation has an accuracy of 20 significant figures

    On the inevitability of the consistency operator

    Full text link
    We examine recursive monotonic functions on the Lindenbaum algebra of EA\mathsf{EA}. We prove that no such function sends every consistent φ\varphi to a sentence with deductive strength strictly between φ\varphi and (φCon(φ))(\varphi\wedge\mathsf{Con}(\varphi)). We generalize this result to iterates of consistency into the effective transfinite. We then prove that for any recursive monotonic function ff, if there is an iterate of Con\mathsf{Con} that bounds ff everywhere, then ff must be somewhere equal to an iterate of Con\mathsf{Con}

    On Stochastic Subgradient Mirror-Descent Algorithm with Weighted Averaging

    Full text link
    This paper considers stochastic subgradient mirror-descent method for solving constrained convex minimization problems. In particular, a stochastic subgradient mirror-descent method with weighted iterate-averaging is investigated and its per-iterate convergence rate is analyzed. The novel part of the approach is in the choice of weights that are used to construct the averages. Through the use of these weighted averages, we show that the known optimal rates can be obtained with simpler algorithms than those currently existing in the literature. Specifically, by suitably choosing the stepsize values, one can obtain the rate of the order 1/k1/k for strongly convex functions, and the rate 1/k1/\sqrt{k} for general convex functions (not necessarily differentiable). Furthermore, for the latter case, it is shown that a stochastic subgradient mirror-descent with iterate averaging converges (along a subsequence) to an optimal solution, almost surely, even with the stepsize of the form 1/1+k1/\sqrt{1+k}, which was not previously known. The stepsize choices that achieve the best rates are those proposed by Paul Tseng for acceleration of proximal gradient methods

    Algebraic entropy and the space of initial values for discrete dynamical systems

    Full text link
    A method to calculate the algebraic entropy of a mapping which can be lifted to an isomorphism of a suitable rational surfaces (the space of initial values) are presented. It is shown that the degree of the nnth iterate of such a mapping is given by its action on the Picard group of the space of initial values. It is also shown that the degree of the nnth iterate of every Painlev\'e equation in sakai's list is at most O(n2)O(n^2) and therefore its algebraic entropy is zero.Comment: 10 pages, pLatex fil
    corecore