27,085 research outputs found

    On the inverse problem of source reconstruction from coherence measurements

    Full text link
    We consider an inverse source problem for partially coherent light propagating in the Fresnel regime. The data is the coherence of the field measured away from the source. The reconstruction is based on a minimum residue formulation, which uses the authors' recent closed-form approximation formula for the coherence of the propagated field. The developed algorithms require a small data sample for convergence and yield stable inversion by exploiting information in the coherence as opposed to intensity-only measurements. Examples with both simulated and experimental data demonstrate the ability of the proposed approach to simultaneously recover complex sources in different planes transverse to the direction of propagation

    Beyond Convexity: Stochastic Quasi-Convex Optimization

    Full text link
    Stochastic convex optimization is a basic and well studied primitive in machine learning. It is well known that convex and Lipschitz functions can be minimized efficiently using Stochastic Gradient Descent (SGD). The Normalized Gradient Descent (NGD) algorithm, is an adaptation of Gradient Descent, which updates according to the direction of the gradients, rather than the gradients themselves. In this paper we analyze a stochastic version of NGD and prove its convergence to a global minimum for a wider class of functions: we require the functions to be quasi-convex and locally-Lipschitz. Quasi-convexity broadens the con- cept of unimodality to multidimensions and allows for certain types of saddle points, which are a known hurdle for first-order optimization methods such as gradient descent. Locally-Lipschitz functions are only required to be Lipschitz in a small region around the optimum. This assumption circumvents gradient explosion, which is another known hurdle for gradient descent variants. Interestingly, unlike the vanilla SGD algorithm, the stochastic normalized gradient descent algorithm provably requires a minimal minibatch size

    DANTE: Deep AlterNations for Training nEural networks

    Full text link
    We present DANTE, a novel method for training neural networks using the alternating minimization principle. DANTE provides an alternate perspective to traditional gradient-based backpropagation techniques commonly used to train deep networks. It utilizes an adaptation of quasi-convexity to cast training a neural network as a bi-quasi-convex optimization problem. We show that for neural network configurations with both differentiable (e.g. sigmoid) and non-differentiable (e.g. ReLU) activation functions, we can perform the alternations effectively in this formulation. DANTE can also be extended to networks with multiple hidden layers. In experiments on standard datasets, neural networks trained using the proposed method were found to be promising and competitive to traditional backpropagation techniques, both in terms of quality of the solution, as well as training speed.Comment: 19 page
    corecore