3,373 research outputs found
Renormalization and Computation II: Time Cut-off and the Halting Problem
This is the second installment to the project initiated in [Ma3]. In the
first Part, I argued that both philosophy and technique of the perturbative
renormalization in quantum field theory could be meaningfully transplanted to
the theory of computation, and sketched several contexts supporting this view.
In this second part, I address some of the issues raised in [Ma3] and provide
their development in three contexts: a categorification of the algorithmic
computations; time cut--off and Anytime Algorithms; and finally, a Hopf algebra
renormalization of the Halting Problem.Comment: 28 page
Computing with and without arbitrary large numbers
In the study of random access machines (RAMs) it has been shown that the
availability of an extra input integer, having no special properties other than
being sufficiently large, is enough to reduce the computational complexity of
some problems. However, this has only been shown so far for specific problems.
We provide a characterization of the power of such extra inputs for general
problems. To do so, we first correct a classical result by Simon and Szegedy
(1992) as well as one by Simon (1981). In the former we show mistakes in the
proof and correct these by an entirely new construction, with no great change
to the results. In the latter, the original proof direction stands with only
minor modifications, but the new results are far stronger than those of Simon
(1981). In both cases, the new constructions provide the theoretical tools
required to characterize the power of arbitrary large numbers.Comment: 12 pages (main text) + 30 pages (appendices), 1 figure. Extended
abstract. The full paper was presented at TAMC 2013. (Reference given is for
the paper version, as it appears in the proceedings.
Two-Way Automata Making Choices Only at the Endmarkers
The question of the state-size cost for simulation of two-way
nondeterministic automata (2NFAs) by two-way deterministic automata (2DFAs) was
raised in 1978 and, despite many attempts, it is still open. Subsequently, the
problem was attacked by restricting the power of 2DFAs (e.g., using a
restricted input head movement) to the degree for which it was already possible
to derive some exponential gaps between the weaker model and the standard
2NFAs. Here we use an opposite approach, increasing the power of 2DFAs to the
degree for which it is still possible to obtain a subexponential conversion
from the stronger model to the standard 2DFAs. In particular, it turns out that
subexponential conversion is possible for two-way automata that make
nondeterministic choices only when the input head scans one of the input tape
endmarkers. However, there is no restriction on the input head movement. This
implies that an exponential gap between 2NFAs and 2DFAs can be obtained only
for unrestricted 2NFAs using capabilities beyond the proposed new model. As an
additional bonus, conversion into a machine for the complement of the original
language is polynomial in this model. The same holds for making such machines
self-verifying, halting, or unambiguous. Finally, any superpolynomial lower
bound for the simulation of such machines by standard 2DFAs would imply LNL.
In the same way, the alternating version of these machines is related to L =?
NL =? P, the classical computational complexity problems.Comment: 23 page
Discounting in LTL
In recent years, there is growing need and interest in formalizing and
reasoning about the quality of software and hardware systems. As opposed to
traditional verification, where one handles the question of whether a system
satisfies, or not, a given specification, reasoning about quality addresses the
question of \emph{how well} the system satisfies the specification. One
direction in this effort is to refine the "eventually" operators of temporal
logic to {\em discounting operators}: the satisfaction value of a specification
is a value in , where the longer it takes to fulfill eventuality
requirements, the smaller the satisfaction value is.
In this paper we introduce an augmentation by discounting of Linear Temporal
Logic (LTL), and study it, as well as its combination with propositional
quality operators. We show that one can augment LTL with an arbitrary set of
discounting functions, while preserving the decidability of the model-checking
problem. Further augmenting the logic with unary propositional quality
operators preserves decidability, whereas adding an average-operator makes some
problems undecidable. We also discuss the complexity of the problem, as well as
various extensions
- …