29,200 research outputs found

    Condorcet-Consistent and Approximately Strategyproof Tournament Rules

    Get PDF
    We consider the manipulability of tournament rules for round-robin tournaments of nn competitors. Specifically, nn competitors are competing for a prize, and a tournament rule rr maps the result of all (n2)\binom{n}{2} pairwise matches (called a tournament, TT) to a distribution over winners. Rule rr is Condorcet-consistent if whenever ii wins all n1n-1 of her matches, rr selects ii with probability 11. We consider strategic manipulation of tournaments where player jj might throw their match to player ii in order to increase the likelihood that one of them wins the tournament. Regardless of the reason why jj chooses to do this, the potential for manipulation exists as long as Pr[r(T)=i]\Pr[r(T) = i] increases by more than Pr[r(T)=j]\Pr[r(T) = j] decreases. Unfortunately, it is known that every Condorcet-consistent rule is manipulable (Altman and Kleinberg). In this work, we address the question of how manipulable Condorcet-consistent rules must necessarily be - by trying to minimize the difference between the increase in Pr[r(T)=i]\Pr[r(T) = i] and decrease in Pr[r(T)=j]\Pr[r(T) = j] for any potential manipulating pair. We show that every Condorcet-consistent rule is in fact 1/31/3-manipulable, and that selecting a winner according to a random single elimination bracket is not α\alpha-manipulable for any α>1/3\alpha > 1/3. We also show that many previously studied tournament formats are all 1/21/2-manipulable, and the popular class of Copeland rules (any rule that selects a player with the most wins) are all in fact 11-manipulable, the worst possible. Finally, we consider extensions to match-fixing among sets of more than two players.Comment: 20 page

    Making proofs without Modus Ponens: An introduction to the combinatorics and complexity of cut elimination

    Full text link
    This paper is intended to provide an introduction to cut elimination which is accessible to a broad mathematical audience. Gentzen's cut elimination theorem is not as well known as it deserves to be, and it is tied to a lot of interesting mathematical structure. In particular we try to indicate some dynamical and combinatorial aspects of cut elimination, as well as its connections to complexity theory. We discuss two concrete examples where one can see the structure of short proofs with cuts, one concerning feasible numbers and the other concerning "bounded mean oscillation" from real analysis

    Simple and Efficient Local Codes for Distributed Stable Network Construction

    Full text link
    In this work, we study protocols so that populations of distributed processes can construct networks. In order to highlight the basic principles of distributed network construction we keep the model minimal in all respects. In particular, we assume finite-state processes that all begin from the same initial state and all execute the same protocol (i.e. the system is homogeneous). Moreover, we assume pairwise interactions between the processes that are scheduled by an adversary. The only constraint on the adversary scheduler is that it must be fair. In order to allow processes to construct networks, we let them activate and deactivate their pairwise connections. When two processes interact, the protocol takes as input the states of the processes and the state of the their connection and updates all of them. Initially all connections are inactive and the goal is for the processes, after interacting and activating/deactivating connections for a while, to end up with a desired stable network. We give protocols (optimal in some cases) and lower bounds for several basic network construction problems such as spanning line, spanning ring, spanning star, and regular network. We provide proofs of correctness for all of our protocols and analyze the expected time to convergence of most of them under a uniform random scheduler that selects the next pair of interacting processes uniformly at random from all such pairs. Finally, we prove several universality results by presenting generic protocols that are capable of simulating a Turing Machine (TM) and exploiting it in order to construct a large class of networks.Comment: 43 pages, 7 figure

    Transfer Function Synthesis without Quantifier Elimination

    Get PDF
    Traditionally, transfer functions have been designed manually for each operation in a program, instruction by instruction. In such a setting, a transfer function describes the semantics of a single instruction, detailing how a given abstract input state is mapped to an abstract output state. The net effect of a sequence of instructions, a basic block, can then be calculated by composing the transfer functions of the constituent instructions. However, precision can be improved by applying a single transfer function that captures the semantics of the block as a whole. Since blocks are program-dependent, this approach necessitates automation. There has thus been growing interest in computing transfer functions automatically, most notably using techniques based on quantifier elimination. Although conceptually elegant, quantifier elimination inevitably induces a computational bottleneck, which limits the applicability of these methods to small blocks. This paper contributes a method for calculating transfer functions that finesses quantifier elimination altogether, and can thus be seen as a response to this problem. The practicality of the method is demonstrated by generating transfer functions for input and output states that are described by linear template constraints, which include intervals and octagons.Comment: 37 pages, extended version of ESOP 2011 pape

    Inactivation Decoding of LT and Raptor Codes: Analysis and Code Design

    Get PDF
    In this paper we analyze LT and Raptor codes under inactivation decoding. A first order analysis is introduced, which provides the expected number of inactivations for an LT code, as a function of the output distribution, the number of input symbols and the decoding overhead. The analysis is then extended to the calculation of the distribution of the number of inactivations. In both cases, random inactivation is assumed. The developed analytical tools are then exploited to design LT and Raptor codes, enabling a tight control on the decoding complexity vs. failure probability trade-off. The accuracy of the approach is confirmed by numerical simulations.Comment: Accepted for publication in IEEE Transactions on Communication
    corecore