928 research outputs found

    Pseudorandom Generators for Width-3 Branching Programs

    Full text link
    We construct pseudorandom generators of seed length O~(log⁥(n)⋅log⁥(1/Ï”))\tilde{O}(\log(n)\cdot \log(1/\epsilon)) that Ï”\epsilon-fool ordered read-once branching programs (ROBPs) of width 33 and length nn. For unordered ROBPs, we construct pseudorandom generators with seed length O~(log⁥(n)⋅poly(1/Ï”))\tilde{O}(\log(n) \cdot \mathrm{poly}(1/\epsilon)). This is the first improvement for pseudorandom generators fooling width 33 ROBPs since the work of Nisan [Combinatorica, 1992]. Our constructions are based on the `iterated milder restrictions' approach of Gopalan et al. [FOCS, 2012] (which further extends the Ajtai-Wigderson framework [FOCS, 1985]), combined with the INW-generator [STOC, 1994] at the last step (as analyzed by Braverman et al. [SICOMP, 2014]). For the unordered case, we combine iterated milder restrictions with the generator of Chattopadhyay et al. [CCC, 2018]. Two conceptual ideas that play an important role in our analysis are: (1) A relabeling technique allowing us to analyze a relabeled version of the given branching program, which turns out to be much easier. (2) Treating the number of colliding layers in a branching program as a progress measure and showing that it reduces significantly under pseudorandom restrictions. In addition, we achieve nearly optimal seed-length O~(log⁥(n/Ï”))\tilde{O}(\log(n/\epsilon)) for the classes of: (1) read-once polynomials on nn variables, (2) locally-monotone ROBPs of length nn and width 33 (generalizing read-once CNFs and DNFs), and (3) constant-width ROBPs of length nn having a layer of width 22 in every consecutive polylog⁥(n)\mathrm{poly}\log(n) layers.Comment: 51 page

    A Mathematical Approach to the Study of the United States Code

    Full text link
    The United States Code (Code) is a document containing over 22 million words that represents a large and important source of Federal statutory law. Scholars and policy advocates often discuss the direction and magnitude of changes in various aspects of the Code. However, few have mathematically formalized the notions behind these discussions or directly measured the resulting representations. This paper addresses the current state of the literature in two ways. First, we formalize a representation of the United States Code as the union of a hierarchical network and a citation network over vertices containing the language of the Code. This representation reflects the fact that the Code is a hierarchically organized document containing language and explicit citations between provisions. Second, we use this formalization to measure aspects of the Code as codified in October 2008, November 2009, and March 2010. These measurements allow for a characterization of the actual changes in the Code over time. Our findings indicate that in the recent past, the Code has grown in its amount of structure, interdependence, and language.Comment: 5 pages, 6 figures, 2 tables

    Visual search in ecological and non-ecological displays: Evidence for a non-monotonic effect of complexity on performance

    Get PDF
    Copyright @ 2013 PLoSThis article has been made available through the Brunel Open Access Publishing Fund.Considerable research has been carried out on visual search, with single or multiple targets. However, most studies have used artificial stimuli with low ecological validity. In addition, little is known about the effects of target complexity and expertise in visual search. Here, we investigate visual search in three conditions of complexity (detecting a king, detecting a check, and detecting a checkmate) with chess players of two levels of expertise (novices and club players). Results show that the influence of target complexity depends on level of structure of the visual display. Different functional relationships were found between artificial (random chess positions) and ecologically valid (game positions) stimuli: With artificial, but not with ecologically valid stimuli, a “pop out” effect was present when a target was visually more complex than distractors but could be captured by a memory chunk. This suggests that caution should be exercised when generalising from experiments using artificial stimuli with low ecological validity to real-life stimuli.This study is funded by Brunel University and the article is made available through the Brunel Open Access Publishing Fund

    Pseudorandomness for Regular Branching Programs via Fourier Analysis

    Full text link
    We present an explicit pseudorandom generator for oblivious, read-once, permutation branching programs of constant width that can read their input bits in any order. The seed length is O(log⁥2n)O(\log^2 n), where nn is the length of the branching program. The previous best seed length known for this model was n1/2+o(1)n^{1/2+o(1)}, which follows as a special case of a generator due to Impagliazzo, Meka, and Zuckerman (FOCS 2012) (which gives a seed length of s1/2+o(1)s^{1/2+o(1)} for arbitrary branching programs of size ss). Our techniques also give seed length n1/2+o(1)n^{1/2+o(1)} for general oblivious, read-once branching programs of width 2no(1)2^{n^{o(1)}}, which is incomparable to the results of Impagliazzo et al.Our pseudorandom generator is similar to the one used by Gopalan et al. (FOCS 2012) for read-once CNFs, but the analysis is quite different; ours is based on Fourier analysis of branching programs. In particular, we show that an oblivious, read-once, regular branching program of width ww has Fourier mass at most (2w2)k(2w^2)^k at level kk, independent of the length of the program.Comment: RANDOM 201

    A general lower bound for collaborative tree exploration

    Full text link
    We consider collaborative graph exploration with a set of kk agents. All agents start at a common vertex of an initially unknown graph and need to collectively visit all other vertices. We assume agents are deterministic, vertices are distinguishable, moves are simultaneous, and we allow agents to communicate globally. For this setting, we give the first non-trivial lower bounds that bridge the gap between small (k≀nk \leq \sqrt n) and large (k≄nk \geq n) teams of agents. Remarkably, our bounds tightly connect to existing results in both domains. First, we significantly extend a lower bound of Ω(log⁥k/log⁥log⁥k)\Omega(\log k / \log\log k) by Dynia et al. on the competitive ratio of a collaborative tree exploration strategy to the range k≀nlog⁥cnk \leq n \log^c n for any c∈Nc \in \mathbb{N}. Second, we provide a tight lower bound on the number of agents needed for any competitive exploration algorithm. In particular, we show that any collaborative tree exploration algorithm with k=Dn1+o(1)k = Dn^{1+o(1)} agents has a competitive ratio of ω(1)\omega(1), while Dereniowski et al. gave an algorithm with k=Dn1+Δk = Dn^{1+\varepsilon} agents and competitive ratio O(1)O(1), for any Δ>0\varepsilon > 0 and with DD denoting the diameter of the graph. Lastly, we show that, for any exploration algorithm using k=nk = n agents, there exist trees of arbitrarily large height DD that require Ω(D2)\Omega(D^2) rounds, and we provide a simple algorithm that matches this bound for all trees

    Antibiotic Resistance Patterns of Bacterial Isolates from Blood in San Francisco County, California, 1996-1999

    Get PDF
    Countywide antibiotic resistance patterns may provide additional information from that obtained from national sampling or individual hospitals. We reviewed susceptibility patterns of selected bacterial strains isolated from blood in San Francisco County from January 1996 to March 1999. We found substantial hospital-to-hospital variability in proportional resistance to antibiotics in multiple organisms. This variability was not correlated with hospital indices such as number of intensive care unit or total beds, annual admissions, or average length of stay. We also found a significant increase in methicillin-resistant Staphylococcus aureus, vancomycin-resistant Enterococcus, and proportional resistance to multiple antipseudomonal antibiotics. We describe the utility, difficulties, and limitations of countywide surveillance

    Modulus Computational Entropy

    Full text link
    The so-called {\em leakage-chain rule} is a very important tool used in many security proofs. It gives an upper bound on the entropy loss of a random variable XX in case the adversary who having already learned some random variables Z1,
,ZℓZ_{1},\ldots,Z_{\ell} correlated with XX, obtains some further information Zℓ+1Z_{\ell+1} about XX. Analogously to the information-theoretic case, one might expect that also for the \emph{computational} variants of entropy the loss depends only on the actual leakage, i.e. on Zℓ+1Z_{\ell+1}. Surprisingly, Krenn et al.\ have shown recently that for the most commonly used definitions of computational entropy this holds only if the computational quality of the entropy deteriorates exponentially in ∣(Z1,
,Zℓ)∣|(Z_{1},\ldots,Z_{\ell})|. This means that the current standard definitions of computational entropy do not allow to fully capture leakage that occurred "in the past", which severely limits the applicability of this notion. As a remedy for this problem we propose a slightly stronger definition of the computational entropy, which we call the \emph{modulus computational entropy}, and use it as a technical tool that allows us to prove a desired chain rule that depends only on the actual leakage and not on its history. Moreover, we show that the modulus computational entropy unifies other,sometimes seemingly unrelated, notions already studied in the literature in the context of information leakage and chain rules. Our results indicate that the modulus entropy is, up to now, the weakest restriction that guarantees that the chain rule for the computational entropy works. As an example of application we demonstrate a few interesting cases where our restricted definition is fulfilled and the chain rule holds.Comment: Accepted at ICTS 201
    • 

    corecore