23,921 research outputs found
Logahedra: A new weakly relational domain
Weakly relational numeric domains express restricted classes of linear inequalities that strike a balance between what can be described and what can be efficiently computed. Popular weakly relational domains such as bounded differences and octagons have found application in model checking and abstract interpretation. This paper introduces logahedra, which are more expressiveness than octagons, but less expressive than arbitrary systems of two variable per inequality constraints. Logahedra allow coefficients of inequalities to be powers of two whilst retaining many of the desirable algorithmic properties of octagons
Spatial discretization of partial differential equations with integrals
We consider the problem of constructing spatial finite difference
approximations on a fixed, arbitrary grid, which have analogues of any number
of integrals of the partial differential equation and of some of its
symmetries. A basis for the space of of such difference operators is
constructed; most cases of interest involve a single such basis element. (The
``Arakawa'' Jacobian is such an element.) We show how the topology of the grid
affects the complexity of the operators.Comment: 24 pages, LaTeX sourc
Group Invariance, Stability to Deformations, and Complexity of Deep Convolutional Representations
The success of deep convolutional architectures is often attributed in part
to their ability to learn multiscale and invariant representations of natural
signals. However, a precise study of these properties and how they affect
learning guarantees is still missing. In this paper, we consider deep
convolutional representations of signals; we study their invariance to
translations and to more general groups of transformations, their stability to
the action of diffeomorphisms, and their ability to preserve signal
information. This analysis is carried by introducing a multilayer kernel based
on convolutional kernel networks and by studying the geometry induced by the
kernel mapping. We then characterize the corresponding reproducing kernel
Hilbert space (RKHS), showing that it contains a large class of convolutional
neural networks with homogeneous activation functions. This analysis allows us
to separate data representation from learning, and to provide a canonical
measure of model complexity, the RKHS norm, which controls both stability and
generalization of any learned model. In addition to models in the constructed
RKHS, our stability analysis also applies to convolutional networks with
generic activations such as rectified linear units, and we discuss its
relationship with recent generalization bounds based on spectral norms
Intensional Models for the Theory of Types
In this paper we define intensional models for the classical theory of types,
thus arriving at an intensional type logic ITL. Intensional models generalize
Henkin's general models and have a natural definition. As a class they do not
validate the axiom of Extensionality. We give a cut-free sequent calculus for
type theory and show completeness of this calculus with respect to the class of
intensional models via a model existence theorem. After this we turn our
attention to applications. Firstly, it is argued that, since ITL is truly
intensional, it can be used to model ascriptions of propositional attitude
without predicting logical omniscience. In order to illustrate this a small
fragment of English is defined and provided with an ITL semantics. Secondly, it
is shown that ITL models contain certain objects that can be identified with
possible worlds. Essential elements of modal logic become available within
classical type theory once the axiom of Extensionality is given up.Comment: 25 page
- …