1,591 research outputs found
Kolmogorov complexity and the asymptotic bound for error-correcting codes
The set of all error-correcting block codes over a fixed alphabet with q letters determines a recursively enumerable set of rational points in the unit square with coordinates (R,δ):= (relative transmission rate, relative minimal distance). Limit points of this set form a closed subset, defined by R≤αq(δ), where αq(δ) is a continuous decreasing function called the asymptotic bound. Its existence was proved by the first-named author in 1981, but no approaches to the computation of this function are known, and in it was even suggested that this function might be uncomputable in the sense of constructive analysis.
In this note we show that the asymptotic bound becomes computable with the assistance of an oracle producing codes in the order of their growing Kolmogorov complexity. Moreover, a natural partition function involving complexity allows us to interpret the asymptotic bound as a curve dividing two different thermodynamic phases of codes
Complexity vs Energy: Theory of Computation and Theoretical Physics
This paper is a survey dedicated to the analogy between the notions of {\it
complexity} in theoretical computer science and {\it energy} in physics. This
analogy is not metaphorical: I describe three precise mathematical contexts,
suggested recently, in which mathematics related to (un)computability is
inspired by and to a degree reproduces formalisms of statistical physics and
quantum field theory.Comment: 23 pages. Talk at the satellite conference to ECM 2012, "QQQ Algebra,
Geometry, Information", Tallinn, July 9-12, 201
Principles and Parameters: a coding theory perspective
We propose an approach to Longobardi's parametric comparison method (PCM) via
the theory of error-correcting codes. One associates to a collection of
languages to be analyzed with the PCM a binary (or ternary) code with one code
words for each language in the family and each word consisting of the binary
values of the syntactic parameters of the language, with the ternary case
allowing for an additional parameter state that takes into account phenomena of
entailment of parameters. The code parameters of the resulting code can be
compared with some classical bounds in coding theory: the asymptotic bound, the
Gilbert-Varshamov bound, etc. The position of the code parameters with respect
to some of these bounds provides quantitative information on the variability of
syntactic parameters within and across historical-linguistic families. While
computations carried out for languages belonging to the same family yield codes
below the GV curve, comparisons across different historical families can give
examples of isolated codes lying above the asymptotic bound.Comment: 11 pages, LaTe
Syntactic Structures and Code Parameters
We assign binary and ternary error-correcting codes to the data of syntactic
structures of world languages and we study the distribution of code points in
the space of code parameters. We show that, while most codes populate the lower
region approximating a superposition of Thomae functions, there is a
substantial presence of codes above the Gilbert-Varshamov bound and even above
the asymptotic bound and the Plotkin bound. We investigate the dynamics induced
on the space of code parameters by spin glass models of language change, and
show that, in the presence of entailment relations between syntactic parameters
the dynamics can sometimes improve the code. For large sets of languages and
syntactic data, one can gain information on the spin glass dynamics from the
induced dynamics in the space of code parameters.Comment: 14 pages, LaTeX, 12 png figure
Zipf's law and L. Levin's probability distributions
Zipf's law in its basic incarnation is an empirical probability distribution
governing the frequency of usage of words in a language. As Terence Tao
recently remarked, it still lacks a convincing and satisfactory mathematical
explanation.
In this paper I suggest that at least in certain situations, Zipf's law can
be explained as a special case of the a priori distribution introduced and
studied by L. Levin. The Zipf ranking corresponding to diminishing probability
appears then as the ordering determined by the growing Kolmogorov complexity.
One argument justifying this assertion is the appeal to a recent
interpretation by Yu. Manin and M. Marcolli of asymptotic bounds for
error--correcting codes in terms of phase transition. In the respective
partition function, Kolmogorov complexity of a code plays the role of its
energy.
This version contains minor corrections and additions.Comment: 19 page
Syntactic Parameters and a Coding Theory Perspective on Entropy and Complexity of Language Families
We present a simple computational approach to assigning a measure of complexity and information/entropy to families of natural languages, based on syntactic parameters and the theory of error correcting codes. We associate to each language a binary string of syntactic parameters and to a language family a binary code, with code words the binary string associated to each language. We then evaluate the code parameters (rate and relative minimum distance) and the position of the parameters with respect to the asymptotic bound of error correcting codes and the Gilbert–Varshamov bound. These bounds are, respectively, related to the Kolmogorov complexity and the Shannon entropy of the code and this gives us a computationally simple way to obtain estimates on the complexity and information, not of individual languages but of language families. This notion of complexity is related, from the linguistic point of view to the degree of variability of syntactic parameter across languages belonging to the same (historical) family
On the Combinatorial Version of the Slepian-Wolf Problem
We study the following combinatorial version of the Slepian-Wolf coding
scheme. Two isolated Senders are given binary strings and respectively;
the length of each string is equal to , and the Hamming distance between the
strings is at most . The Senders compress their strings and
communicate the results to the Receiver. Then the Receiver must reconstruct
both strings and . The aim is to minimise the lengths of the transmitted
messages.
For an asymmetric variant of this problem (where one of the Senders transmits
the input string to the Receiver without compression) with deterministic
encoding a nontrivial lower bound was found by A.Orlitsky and K.Viswanathany.
In our paper we prove a new lower bound for the schemes with syndrome coding,
where at least one of the Senders uses linear encoding of the input string.
For the combinatorial Slepian-Wolf problem with randomized encoding the
theoretical optimum of communication complexity was recently found by the first
author, though effective protocols with optimal lengths of messages remained
unknown. We close this gap and present a polynomial time randomized protocol
that achieves the optimal communication complexity.Comment: 20 pages, 14 figures. Accepted to IEEE Transactions on Information
Theory (June 2018
- …