20,313 research outputs found
A Continuous-Discontinuous Second-Order Transition in the Satisfiability of Random Horn-SAT Formulas
We compute the probability of satisfiability of a class of random Horn-SAT
formulae, motivated by a connection with the nonemptiness problem of finite
tree automata. In particular, when the maximum clause length is 3, this model
displays a curve in its parameter space along which the probability of
satisfiability is discontinuous, ending in a second-order phase transition
where it becomes continuous. This is the first case in which a phase transition
of this type has been rigorously established for a random constraint
satisfaction problem
Time Resolution Dependence of Information Measures for Spiking Neurons: Atoms, Scaling, and Universality
The mutual information between stimulus and spike-train response is commonly
used to monitor neural coding efficiency, but neuronal computation broadly
conceived requires more refined and targeted information measures of
input-output joint processes. A first step towards that larger goal is to
develop information measures for individual output processes, including
information generation (entropy rate), stored information (statistical
complexity), predictable information (excess entropy), and active information
accumulation (bound information rate). We calculate these for spike trains
generated by a variety of noise-driven integrate-and-fire neurons as a function
of time resolution and for alternating renewal processes. We show that their
time-resolution dependence reveals coarse-grained structural properties of
interspike interval statistics; e.g., -entropy rates that diverge less
quickly than the firing rate indicate interspike interval correlations. We also
find evidence that the excess entropy and regularized statistical complexity of
different types of integrate-and-fire neurons are universal in the
continuous-time limit in the sense that they do not depend on mechanism
details. This suggests a surprising simplicity in the spike trains generated by
these model neurons. Interestingly, neurons with gamma-distributed ISIs and
neurons whose spike trains are alternating renewal processes do not fall into
the same universality class. These results lead to two conclusions. First, the
dependence of information measures on time resolution reveals mechanistic
details about spike train generation. Second, information measures can be used
as model selection tools for analyzing spike train processes.Comment: 20 pages, 6 figures;
http://csc.ucdavis.edu/~cmg/compmech/pubs/trdctim.ht
Accurate and efficient evaluation of the a posteriori error estimator in the reduced basis method
The reduced basis method is a model reduction technique yielding substantial
savings of computational time when a solution to a parametrized equation has to
be computed for many values of the parameter. Certification of the
approximation is possible by means of an a posteriori error bound. Under
appropriate assumptions, this error bound is computed with an algorithm of
complexity independent of the size of the full problem. In practice, the
evaluation of the error bound can become very sensitive to round-off errors. We
propose herein an explanation of this fact. A first remedy has been proposed in
[F. Casenave, Accurate \textit{a posteriori} error evaluation in the reduced
basis method. \textit{C. R. Math. Acad. Sci. Paris} \textbf{350} (2012)
539--542.]. Herein, we improve this remedy by proposing a new approximation of
the error bound using the Empirical Interpolation Method (EIM). This method
achieves higher levels of accuracy and requires potentially less
precomputations than the usual formula. A version of the EIM stabilized with
respect to round-off errors is also derived. The method is illustrated on a
simple one-dimensional diffusion problem and a three-dimensional acoustic
scattering problem solved by a boundary element method.Comment: 26 pages, 10 figures. ESAIM: Mathematical Modelling and Numerical
Analysis, 201
From Small Space to Small Width in Resolution
In 2003, Atserias and Dalmau resolved a major open question about the
resolution proof system by establishing that the space complexity of CNF
formulas is always an upper bound on the width needed to refute them. Their
proof is beautiful but somewhat mysterious in that it relies heavily on tools
from finite model theory. We give an alternative, completely elementary proof
that works by simple syntactic manipulations of resolution refutations. As a
by-product, we develop a "black-box" technique for proving space lower bounds
via a "static" complexity measure that works against any resolution
refutation---previous techniques have been inherently adaptive. We conclude by
showing that the related question for polynomial calculus (i.e., whether space
is an upper bound on degree) seems unlikely to be resolvable by similar
methods
Space proof complexity for random 3-CNFs
We investigate the space complexity of refuting 3-CNFs in Resolution and algebraic systems. We prove that every Polynomial Calculus with Resolution refutation of a random 3-CNF φ in n variables requires, with high probability, distinct monomials to be kept simultaneously in memory. The same construction also proves that every Resolution refutation of φ requires, with high probability, clauses each of width to be kept at the same time in memory. This gives a lower bound for the total space needed in Resolution to refute φ. These results are best possible (up to a constant factor) and answer questions about space complexity of 3-CNFs
- …