5,115 research outputs found

    Fast derivatives of likelihood functionals for ODE based models using adjoint-state method

    Full text link
    We consider time series data modeled by ordinary differential equations (ODEs), widespread models in physics, chemistry, biology and science in general. The sensitivity analysis of such dynamical systems usually requires calculation of various derivatives with respect to the model parameters. We employ the adjoint state method (ASM) for efficient computation of the first and the second derivatives of likelihood functionals constrained by ODEs with respect to the parameters of the underlying ODE model. Essentially, the gradient can be computed with a cost (measured by model evaluations) that is independent of the number of the ODE model parameters and the Hessian with a linear cost in the number of the parameters instead of the quadratic one. The sensitivity analysis becomes feasible even if the parametric space is high-dimensional. The main contributions are derivation and rigorous analysis of the ASM in the statistical context, when the discrete data are coupled with the continuous ODE model. Further, we present a highly optimized implementation of the results and its benchmarks on a number of problems. The results are directly applicable in (e.g.) maximum-likelihood estimation or Bayesian sampling of ODE based statistical models, allowing for faster, more stable estimation of parameters of the underlying ODE model.Comment: 5 figure

    Learning to Prune Deep Neural Networks via Layer-wise Optimal Brain Surgeon

    Full text link
    How to develop slim and accurate deep neural networks has become crucial for real- world applications, especially for those employed in embedded systems. Though previous work along this research line has shown some promising results, most existing methods either fail to significantly compress a well-trained deep network or require a heavy retraining process for the pruned deep network to re-boost its prediction performance. In this paper, we propose a new layer-wise pruning method for deep neural networks. In our proposed method, parameters of each individual layer are pruned independently based on second order derivatives of a layer-wise error function with respect to the corresponding parameters. We prove that the final prediction performance drop after pruning is bounded by a linear combination of the reconstructed errors caused at each layer. Therefore, there is a guarantee that one only needs to perform a light retraining process on the pruned network to resume its original prediction performance. We conduct extensive experiments on benchmark datasets to demonstrate the effectiveness of our pruning method compared with several state-of-the-art baseline methods

    Delineating Parameter Unidentifiabilities in Complex Models

    Full text link
    Scientists use mathematical modelling to understand and predict the properties of complex physical systems. In highly parameterised models there often exist relationships between parameters over which model predictions are identical, or nearly so. These are known as structural or practical unidentifiabilities, respectively. They are hard to diagnose and make reliable parameter estimation from data impossible. They furthermore imply the existence of an underlying model simplification. We describe a scalable method for detecting unidentifiabilities, and the functional relations defining them, for generic models. This allows for model simplification, and appreciation of which parameters (or functions thereof) cannot be estimated from data. Our algorithm can identify features such as redundant mechanisms and fast timescale subsystems, as well as the regimes in which such approximations are valid. We base our algorithm on a novel quantification of regional parametric sensitivity: multiscale sloppiness. Traditionally, the link between parametric sensitivity and the conditioning of the parameter estimation problem is made locally, through the Fisher Information Matrix. This is valid in the regime of infinitesimal measurement uncertainty. We demonstrate the duality between multiscale sloppiness and the geometry of confidence regions surrounding parameter estimates made where measurement uncertainty is non-negligible. Further theoretical relationships are provided linking multiscale sloppiness to the Likelihood-ratio test. From this, we show that a local sensitivity analysis (as typically done) is insufficient for determining the reliability of parameter estimation, even with simple (non)linear systems. Our algorithm provides a tractable alternative. We finally apply our methods to a large-scale, benchmark Systems Biology model of NF-κ\kappaB, uncovering previously unknown unidentifiabilities

    Shape Hessian for generalized Oseen flow by differentiability of a minimax: A Lagrangian approach

    Get PDF
    summary:The goal of this paper is to compute the shape Hessian for a generalized Oseen problem with nonhomogeneous Dirichlet boundary condition by the velocity method. The incompressibility will be treated by penalty approach. The structure of the shape gradient and shape Hessian with respect to the shape of the variable domain for a given cost functional are established by an application of the Lagrangian method with function space embedding technique

    Parametric Level Set Methods for Inverse Problems

    Full text link
    In this paper, a parametric level set method for reconstruction of obstacles in general inverse problems is considered. General evolution equations for the reconstruction of unknown obstacles are derived in terms of the underlying level set parameters. We show that using the appropriate form of parameterizing the level set function results a significantly lower dimensional problem, which bypasses many difficulties with traditional level set methods, such as regularization, re-initialization and use of signed distance function. Moreover, we show that from a computational point of view, low order representation of the problem paves the path for easier use of Newton and quasi-Newton methods. Specifically for the purposes of this paper, we parameterize the level set function in terms of adaptive compactly supported radial basis functions, which used in the proposed manner provides flexibility in presenting a larger class of shapes with fewer terms. Also they provide a "narrow-banding" advantage which can further reduce the number of active unknowns at each step of the evolution. The performance of the proposed approach is examined in three examples of inverse problems, i.e., electrical resistance tomography, X-ray computed tomography and diffuse optical tomography

    Errors in quantum optimal control and strategy for the search of easily implementable control pulses

    Full text link
    We introduce a new approach to assess the error of control problems we aim to optimize. The method offers a strategy to define new control pulses that are not necessarily optimal but still able to yield an error not larger than some fixed a priori threshold, and therefore provide control pulses that might be more amenable for an experimental implementation. The formalism is applied to an exactly solvable model and to the Landau-Zener model, whose optimal control problem is solvable only numerically. The presented method is of importance for applications where a high degree of controllability of the dynamics of quantum systems is required.Comment: 13 pages, 3 figure

    On the Second-Order Shape Derivative of the Kohn-Vogelius Objective Functional Using the Velocity Method

    Get PDF
    The exterior Bernoulli free boundary problem was studied via shape optimization technique. The problem was reformulated into the minimization of the so-called Kohn-Vogelius objective functional, where two state variables involved satisfy two boundary value problems, separately. The paper focused on solving the second-order shape derivative of the objective functional using the velocity method with nonautonomous velocity fields. This work confirms the classical results of Delfour and Zolésio in relating shape derivatives of functionals using velocity method and perturbation of identity technique
    corecore