42 research outputs found

    Hybrid symbolic-numeric methods in geosciences

    Get PDF
    Modern computing systems, like Mathematica, are blending numeric and symbolic evaluation of many algorithms improving their efficiencies (time, accuracy). In this contribution three toy-examples of the application of hybrid symbolic-numeric computations in geosciences are presented in order to illustrate the features of this advanced technique, namely: ranging GNSS satellites, computing GNSS cycle ambiguities and employing symbolic regression for verifying Kepler’s third law

    Solving polynomial systems via symbolic-numeric reduction to geometric involutive form

    Get PDF
    AbstractWe briefly survey several existing methods for solving polynomial systems with inexact coefficients, then introduce our new symbolic-numeric method which is based on the geometric (Jet) theory of partial differential equations. The method is stable and robust. Numerical experiments illustrate the performance of the new method

    Joint shape and motion estimation from echo-based sensor data

    Get PDF
    2018 Fall.Includes bibliographical references.Given a set of time-series data collected from echo-based ranging sensors, we study the problem of jointly estimating the shape and motion of the target under observation when the sensor positions are also unknown. Using an approach first described by Stuff et al., we model the target as a point configuration in Euclidean space and estimate geometric invariants of the configuration. The geometric invariants allow us to estimate the target shape, from which we can estimate the motion of the target relative to the sensor position. This work will unify the various geometric- invariant based shape and motion estimation literature under a common framework, and extend that framework to include results for passive, bistatic sensor systems

    Tropical algebraic geometry in Maple: A preprocessing algorithm for finding common factors for multivariate polynomials with approximate coefficients

    Get PDF
    AbstractFinding a common factor of two multivariate polynomials with approximate coefficients is a problem in symbolic–numeric computing. Taking a tropical view of this problem leads to efficient preprocessing techniques, applying polyhedral methods to the exact exponents and numerical techniques to the approximate coefficients. With Maple we will illustrate our use of tropical algebraic geometry

    Using Machine Learning to Improve Cylindrical Algebraic Decomposition

    Get PDF
    Cylindrical Algebraic Decomposition (CAD) is a key tool in computational algebraic geometry, best known as a procedure to enable Quantifier Elimination over real-closed fields. However, it has a worst case complexity doubly exponential in the size of the input, which is often encountered in practice. It has been observed that for many problems a change in algorithm settings or problem formulation can cause huge differences in runtime costs, changing problem instances from intractable to easy. A number of heuristics have been developed to help with such choices, but the complicated nature of the geometric relationships involved means these are imperfect and can sometimes make poor choices. We investigate the use of machine learning (specifically support vector machines) to make such choices instead. Machine learning is the process of fitting a computer model to a complex function based on properties learned from measured data. In this paper we apply it in two case studies: the first to select between heuristics for choosing a CAD variable ordering; the second to identify when a CAD problem instance would benefit from Groebner Basis preconditioning. These appear to be the first such applications of machine learning to Symbolic Computation. We demonstrate in both cases that the machine learned choice outperforms human developed heuristics.This work was supported by EPSRC grant EP/J003247/1; the European Union’s Horizon 2020 research and innovation programme under grant agreement No 712689 (SC2); and the China Scholarship Council (CSC)

    Model-based Design of Experiments for Large Dataset

    Get PDF
    The first part of this thesis presents the motivation for adapting ideas and methods from the theory of the model-based optimal design of experiments in the context of Big Data while guarding against different sources of bias. In particular, the key focus is on the issue of guarding against bias from confounders and how to use the theory of the design of experiment and randomization to remove bias depending on the constraints in the design. Starting with A/B experiments, largely used by major Tech Companies in online marketing, the theory of circuits is introduced and an algebraic methods which gives a wide choice of randomization schemes is presented. Furthermore, a robust exchange algorithm to deal with the problem of outliers in a Big Dataset is proposed. The second part is based on a marine insurance use case sponsored by Swiss Re Corporate Solutions, commercial insurance division of the Swiss Re Group. Several temporal disaggregation methods for dealing with time series collected at different time frequencies are reviewed and applied to real data in order to obtain a curated dataset for predicting future losses

    Toward a structural theory of learning algebraic decompositions

    Get PDF
    We propose a framework generalizing several variants of Prony's method and explaining their relations. These methods are suitable for determining the support of linear combinations in particular in vector spaces of functions from evaluations. They are based on suitable sequences of linear maps resp. their matrices and include Hankel and Toeplitz variants of Prony's method for the decomposition of multivariate exponential sums, polynomials (w.r.t. the monomial and Chebyshev bases), Gau fian sums, spherical harmonic sums, taking also into account whether they have their support on an algebraic set

    Joint model of probabilistic-robust (probust) constraints with application to gas network optimization

    Get PDF
    Optimization problems under uncertain conditions abound in many real-life applications. While solution approaches for probabilistic constraints are often developed in case the uncertainties can be assumed to follow a certain probability distribution, robust approaches are usually applied in case solutions are sought that are feasible for all realizations of uncertainties within some predefined uncertainty set. As many applications contain different types of uncertainties that require robust as well as probabilistic treatments, we introduce a class of joint probabilistic/robust constraints. Focusing on complex uncertain gas network optimization problems, we show the relevance of this class of problems for the task of maximizing free booked capacities in an algebraic model for a stationary gas network. We furthermore present approaches for finding their solution. Finally, we study the problem of controlling a transient system that is governed by the wave equation. The task consists in determining controls such that a certain robustness measure remains below some given upper bound with high probability

    Implementation and Evaluation of Algorithmic Skeletons: Parallelisation of Computer Algebra Algorithms

    Get PDF
    This thesis presents design and implementation approaches for the parallel algorithms of computer algebra. We use algorithmic skeletons and also further approaches, like data parallel arithmetic and actors. We have implemented skeletons for divide and conquer algorithms and some special parallel loops, that we call ‘repeated computation with a possibility of premature termination’. We introduce in this thesis a rational data parallel arithmetic. We focus on parallel symbolic computation algorithms, for these algorithms our arithmetic provides a generic parallelisation approach. The implementation is carried out in Eden, a parallel functional programming language based on Haskell. This choice enables us to encode both the skeletons and the programs in the same language. Moreover, it allows us to refrain from using two different languages—one for the implementation and one for the interface—for our implementation of computer algebra algorithms. Further, this thesis presents methods for evaluation and estimation of parallel execution times. We partition the parallel execution time into two components. One of them accounts for the quality of the parallelisation, we call it the ‘parallel penalty’. The other is the sequential execution time. For the estimation, we predict both components separately, using statistical methods. This enables very confident estimations, although using drastically less measurement points than other methods. We have applied both our evaluation and estimation approaches to the parallel programs presented in this thesis. We haven also used existing estimation methods. We developed divide and conquer skeletons for the implementation of fast parallel multiplication. We have implemented the Karatsuba algorithm, Strassen’s matrix multiplication algorithm and the fast Fourier transform. The latter was used to implement polynomial convolution that leads to a further fast multiplication algorithm. Specially for our implementation of Strassen algorithm we have designed and implemented a divide and conquer skeleton basing on actors. We have implemented the parallel fast Fourier transform, and not only did we use new divide and conquer skeletons, but also developed a map-and-transpose skeleton. It enables good parallelisation of the Fourier transform. The parallelisation of Karatsuba multiplication shows a very good performance. We have analysed the parallel penalty of our programs and compared it to the serial fraction—an approach, known from literature. We also performed execution time estimations of our divide and conquer programs. This thesis presents a parallel map+reduce skeleton scheme. It allows us to combine the usual parallel map skeletons, like parMap, farm, workpool, with a premature termination property. We use this to implement the so-called ‘parallel repeated computation’, a special form of a speculative parallel loop. We have implemented two probabilistic primality tests: the Rabin–Miller test and the Jacobi sum test. We parallelised both with our approach. We analysed the task distribution and stated the fitting configurations of the Jacobi sum test. We have shown formally that the Jacobi sum test can be implemented in parallel. Subsequently, we parallelised it, analysed the load balancing issues, and produced an optimisation. The latter enabled a good implementation, as verified using the parallel penalty. We have also estimated the performance of the tests for further input sizes and numbers of processing elements. Parallelisation of the Jacobi sum test and our generic parallelisation scheme for the repeated computation is our original contribution. The data parallel arithmetic was defined not only for integers, which is already known, but also for rationals. We handled the common factors of the numerator or denominator of the fraction with the modulus in a novel manner. This is required to obtain a true multiple-residue arithmetic, a novel result of our research. Using these mathematical advances, we have parallelised the determinant computation using the Gauß elimination. As always, we have performed task distribution analysis and estimation of the parallel execution time of our implementation. A similar computation in Maple emphasised the potential of our approach. Data parallel arithmetic enables parallelisation of entire classes of computer algebra algorithms. Summarising, this thesis presents and thoroughly evaluates new and existing design decisions for high-level parallelisations of computer algebra algorithms
    corecore