172,963 research outputs found
An Algorithmic Approach to Information and Meaning
I will survey some matters of relevance to a philosophical discussion of
information, taking into account developments in algorithmic information theory
(AIT). I will propose that meaning is deep in the sense of Bennett's logical
depth, and that algorithmic probability may provide the stability needed for a
robust algorithmic definition of meaning, one that takes into consideration the
interpretation and the recipient's own knowledge encoded in the story attached
to a message.Comment: preprint reviewed version closer to the version accepted by the
journa
On Universal Prediction and Bayesian Confirmation
The Bayesian framework is a well-studied and successful framework for
inductive reasoning, which includes hypothesis testing and confirmation,
parameter estimation, sequence prediction, classification, and regression. But
standard statistical guidelines for choosing the model class and prior are not
always available or fail, in particular in complex situations. Solomonoff
completed the Bayesian framework by providing a rigorous, unique, formal, and
universal choice for the model class and the prior. We discuss in breadth how
and in which sense universal (non-i.i.d.) sequence prediction solves various
(philosophical) problems of traditional Bayesian sequence prediction. We show
that Solomonoff's model possesses many desirable properties: Strong total and
weak instantaneous bounds, and in contrast to most classical continuous prior
densities has no zero p(oste)rior problem, i.e. can confirm universal
hypotheses, is reparametrization and regrouping invariant, and avoids the
old-evidence and updating problem. It even performs well (actually better) in
non-computable environments.Comment: 24 page
Theory and Techniques for Synthesizing a Family of Graph Algorithms
Although Breadth-First Search (BFS) has several advantages over Depth-First
Search (DFS) its prohibitive space requirements have meant that algorithm
designers often pass it over in favor of DFS. To address this shortcoming, we
introduce a theory of Efficient BFS (EBFS) along with a simple recursive
program schema for carrying out the search. The theory is based on dominance
relations, a long standing technique from the field of search algorithms. We
show how the theory can be used to systematically derive solutions to two graph
algorithms, namely the Single Source Shortest Path problem and the Minimum
Spanning Tree problem. The solutions are found by making small systematic
changes to the derivation, revealing the connections between the two problems
which are often obscured in textbook presentations of them.Comment: In Proceedings SYNT 2012, arXiv:1207.055
p-probabilistic k-anonymous microaggregation for the anonymization of surveys with uncertain participation
We develop a probabilistic variant of k-anonymous microaggregation which we term p-probabilistic resorting to a statistical model of respondent participation in order to aggregate quasi-identifiers in such a manner that k-anonymity is concordantly enforced with a parametric probabilistic guarantee. Succinctly owing the possibility that some respondents may not finally participate, sufficiently larger cells are created striving to satisfy k-anonymity with probability at least p. The microaggregation function is designed before the respondents submit their confidential data. More precisely, a specification of the function is sent to them which they may verify and apply to their quasi-identifying demographic variables prior to submitting the microaggregated data along with the confidential attributes to an authorized repository.
We propose a number of metrics to assess the performance of our probabilistic approach in terms of anonymity and distortion which we proceed to investigate theoretically in depth and empirically with synthetic and standardized data. We stress that in addition to constituting a functional extension of traditional microaggregation, thereby broadening its applicability to the anonymization of statistical databases in a wide variety of contexts, the relaxation of trust assumptions is arguably expected to have a considerable impact on user acceptance and ultimately on data utility through mere availability.Peer ReviewedPostprint (author's final draft
The GIST of Concepts
A unified general theory of human concept learning based on the idea that humans detect invariance patterns in categorical stimuli as a necessary precursor to concept formation is proposed and tested. In GIST (generalized invariance structure theory) invariants are detected via a perturbation mechanism of dimension suppression referred to as dimensional binding. Structural information acquired by this process is stored as a compound memory trace termed an ideotype. Ideotypes inform the subsystems that are responsible for learnability judgments, rule formation, and other types of concept representations. We show that GIST is more general (e.g., it works on continuous, semi-continuous, and binary stimuli) and makes much more accurate predictions than the leading models of concept learning difficulty,such as those based on a complexity reduction principle (e.g., number of mental models,structural invariance, algebraic complexity, and minimal description length) and those based on selective attention and similarity (GCM, ALCOVE, and SUSTAIN). GIST unifies these two key aspects of concept learning and categorization. Empirical evidence from three\ud
experiments corroborates the predictions made by the theory and its core model which we propose as a candidate law of human conceptual behavior
Cake Cutting Algorithms for Piecewise Constant and Piecewise Uniform Valuations
Cake cutting is one of the most fundamental settings in fair division and
mechanism design without money. In this paper, we consider different levels of
three fundamental goals in cake cutting: fairness, Pareto optimality, and
strategyproofness. In particular, we present robust versions of envy-freeness
and proportionality that are not only stronger than their standard
counter-parts but also have less information requirements. We then focus on
cake cutting with piecewise constant valuations and present three desirable
algorithms: CCEA (Controlled Cake Eating Algorithm), MEA (Market Equilibrium
Algorithm) and CSD (Constrained Serial Dictatorship). CCEA is polynomial-time,
robust envy-free, and non-wasteful. It relies on parametric network flows and
recent generalizations of the probabilistic serial algorithm. For the subdomain
of piecewise uniform valuations, we show that it is also group-strategyproof.
Then, we show that there exists an algorithm (MEA) that is polynomial-time,
envy-free, proportional, and Pareto optimal. MEA is based on computing a
market-based equilibrium via a convex program and relies on the results of
Reijnierse and Potters [24] and Devanur et al. [15]. Moreover, we show that MEA
and CCEA are equivalent to mechanism 1 of Chen et. al. [12] for piecewise
uniform valuations. We then present an algorithm CSD and a way to implement it
via randomization that satisfies strategyproofness in expectation, robust
proportionality, and unanimity for piecewise constant valuations. For the case
of two agents, it is robust envy-free, robust proportional, strategyproof, and
polynomial-time. Many of our results extend to more general settings in cake
cutting that allow for variable claims and initial endowments. We also show a
few impossibility results to complement our algorithms.Comment: 39 page
Kolmogorov Complexity in perspective. Part II: Classification, Information Processing and Duality
We survey diverse approaches to the notion of information: from Shannon
entropy to Kolmogorov complexity. Two of the main applications of Kolmogorov
complexity are presented: randomness and classification. The survey is divided
in two parts published in a same volume. Part II is dedicated to the relation
between logic and information system, within the scope of Kolmogorov
algorithmic information theory. We present a recent application of Kolmogorov
complexity: classification using compression, an idea with provocative
implementation by authors such as Bennett, Vitanyi and Cilibrasi. This stresses
how Kolmogorov complexity, besides being a foundation to randomness, is also
related to classification. Another approach to classification is also
considered: the so-called "Google classification". It uses another original and
attractive idea which is connected to the classification using compression and
to Kolmogorov complexity from a conceptual point of view. We present and unify
these different approaches to classification in terms of Bottom-Up versus
Top-Down operational modes, of which we point the fundamental principles and
the underlying duality. We look at the way these two dual modes are used in
different approaches to information system, particularly the relational model
for database introduced by Codd in the 70's. This allows to point out diverse
forms of a fundamental duality. These operational modes are also reinterpreted
in the context of the comprehension schema of axiomatic set theory ZF. This
leads us to develop how Kolmogorov's complexity is linked to intensionality,
abstraction, classification and information system.Comment: 43 page
- …