5,236 research outputs found

    Relative entropy minimizing noisy non-linear neural network to approximate stochastic processes

    Full text link
    A method is provided for designing and training noise-driven recurrent neural networks as models of stochastic processes. The method unifies and generalizes two known separate modeling approaches, Echo State Networks (ESN) and Linear Inverse Modeling (LIM), under the common principle of relative entropy minimization. The power of the new method is demonstrated on a stochastic approximation of the El Nino phenomenon studied in climate research

    Gaussian Process Morphable Models

    Get PDF
    Statistical shape models (SSMs) represent a class of shapes as a normal distribution of point variations, whose parameters are estimated from example shapes. Principal component analysis (PCA) is applied to obtain a low-dimensional representation of the shape variation in terms of the leading principal components. In this paper, we propose a generalization of SSMs, called Gaussian Process Morphable Models (GPMMs). We model the shape variations with a Gaussian process, which we represent using the leading components of its Karhunen-Loeve expansion. To compute the expansion, we make use of an approximation scheme based on the Nystrom method. The resulting model can be seen as a continuous analogon of an SSM. However, while for SSMs the shape variation is restricted to the span of the example data, with GPMMs we can define the shape variation using any Gaussian process. For example, we can build shape models that correspond to classical spline models, and thus do not require any example data. Furthermore, Gaussian processes make it possible to combine different models. For example, an SSM can be extended with a spline model, to obtain a model that incorporates learned shape characteristics, but is flexible enough to explain shapes that cannot be represented by the SSM. We introduce a simple algorithm for fitting a GPMM to a surface or image. This results in a non-rigid registration approach, whose regularization properties are defined by a GPMM. We show how we can obtain different registration schemes,including methods for multi-scale, spatially-varying or hybrid registration, by constructing an appropriate GPMM. As our approach strictly separates modelling from the fitting process, this is all achieved without changes to the fitting algorithm. We show the applicability and versatility of GPMMs on a clinical use case, where the goal is the model-based segmentation of 3D forearm images

    Model Order Reduction for Nonlinear Schr\"odinger Equation

    Full text link
    We apply the proper orthogonal decomposition (POD) to the nonlinear Schr\"odinger (NLS) equation to derive a reduced order model. The NLS equation is discretized in space by finite differences and is solved in time by structure preserving symplectic mid-point rule. A priori error estimates are derived for the POD reduced dynamical system. Numerical results for one and two dimensional NLS equations, coupled NLS equation with soliton solutions show that the low-dimensional approximations obtained by POD reproduce very well the characteristic dynamics of the system, such as preservation of energy and the solutions

    A DEIM Induced CUR Factorization

    Full text link
    We derive a CUR matrix factorization based on the Discrete Empirical Interpolation Method (DEIM). For a given matrix AA, such a factorization provides a low rank approximate decomposition of the form A≈CURA \approx C U R, where CC and RR are subsets of the columns and rows of AA, and UU is constructed to make CURCUR a good approximation. Given a low-rank singular value decomposition A≈VSWTA \approx V S W^T, the DEIM procedure uses VV and WW to select the columns and rows of AA that form CC and RR. Through an error analysis applicable to a general class of CUR factorizations, we show that the accuracy tracks the optimal approximation error within a factor that depends on the conditioning of submatrices of VV and WW. For large-scale problems, VV and WW can be approximated using an incremental QR algorithm that makes one pass through AA. Numerical examples illustrate the favorable performance of the DEIM-CUR method, compared to CUR approximations based on leverage scores
    • …
    corecore