3,668 research outputs found

    Wave polynomials, transmutations and Cauchy's problem for the Klein-Gordon equation

    Full text link
    We prove a completeness result for a class of polynomial solutions of the wave equation called wave polynomials and construct generalized wave polynomials, solutions of the Klein-Gordon equation with a variable coefficient. Using the transmutation (transformation) operators and their recently discovered mapping properties we prove the completeness of the generalized wave polynomials and use them for an explicit construction of the solution of the Cauchy problem for the Klein-Gordon equation. Based on this result we develop a numerical method for solving the Cauchy problem and test its performance.Comment: 31 pages, 8 figures (16 graphs

    Tangential Extremal Principles for Finite and Infinite Systems of Sets, II: Applications to Semi-infinite and Multiobjective Optimization

    Get PDF
    This paper contains selected applications of the new tangential extremal principles and related results developed in Part I to calculus rules for infinite intersections of sets and optimality conditions for problems of semi-infinite programming and multiobjective optimization with countable constraint

    Data-driven Inverse Optimization with Imperfect Information

    Full text link
    In data-driven inverse optimization an observer aims to learn the preferences of an agent who solves a parametric optimization problem depending on an exogenous signal. Thus, the observer seeks the agent's objective function that best explains a historical sequence of signals and corresponding optimal actions. We focus here on situations where the observer has imperfect information, that is, where the agent's true objective function is not contained in the search space of candidate objectives, where the agent suffers from bounded rationality or implementation errors, or where the observed signal-response pairs are corrupted by measurement noise. We formalize this inverse optimization problem as a distributionally robust program minimizing the worst-case risk that the {\em predicted} decision ({\em i.e.}, the decision implied by a particular candidate objective) differs from the agent's {\em actual} response to a random signal. We show that our framework offers rigorous out-of-sample guarantees for different loss functions used to measure prediction errors and that the emerging inverse optimization problems can be exactly reformulated as (or safely approximated by) tractable convex programs when a new suboptimality loss function is used. We show through extensive numerical tests that the proposed distributionally robust approach to inverse optimization attains often better out-of-sample performance than the state-of-the-art approaches

    Precise lower bound on Monster brane boundary entropy

    Full text link
    In this paper we develop further the linear functional method of deriving lower bounds on the boundary entropy of conformal boundary conditions in 1+1 dimensional conformal field theories (CFTs). We show here how to use detailed knowledge of the bulk CFT spectrum. Applying the method to the Monster CFT with c=\bar c=24 we derive a lower bound s > - 3.02 x 10^{-19} on the boundary entropy s=ln g, and find compelling evidence that the optimal bound is s>= 0. We show that all g=1 branes must have the same low-lying boundary spectrum, which matches the spectrum of the known g=1 branes, suggesting that the known examples comprise all possible g=1 branes, and also suggesting that the bound s>= 0 holds not just for critical boundary conditions but for all boundary conditions in the Monster CFT. The same analysis applied to a second bulk CFT -- a certain c=2 Gaussian model -- yields a less strict bound, suggesting that the precise linear functional bound on s for the Monster CFT is exceptional.Comment: 1+18 page

    Super-resolution, Extremal Functions and the Condition Number of Vandermonde Matrices

    Get PDF
    Super-resolution is a fundamental task in imaging, where the goal is to extract fine-grained structure from coarse-grained measurements. Here we are interested in a popular mathematical abstraction of this problem that has been widely studied in the statistics, signal processing and machine learning communities. We exactly resolve the threshold at which noisy super-resolution is possible. In particular, we establish a sharp phase transition for the relationship between the cutoff frequency (mm) and the separation (Δ\Delta). If m>1/Δ+1m > 1/\Delta + 1, our estimator converges to the true values at an inverse polynomial rate in terms of the magnitude of the noise. And when m<(1−ϵ)/Δm < (1-\epsilon) /\Delta no estimator can distinguish between a particular pair of Δ\Delta-separated signals even if the magnitude of the noise is exponentially small. Our results involve making novel connections between {\em extremal functions} and the spectral properties of Vandermonde matrices. We establish a sharp phase transition for their condition number which in turn allows us to give the first noise tolerance bounds for the matrix pencil method. Moreover we show that our methods can be interpreted as giving preconditioners for Vandermonde matrices, and we use this observation to design faster algorithms for super-resolution. We believe that these ideas may have other applications in designing faster algorithms for other basic tasks in signal processing.Comment: 19 page
    • …
    corecore