115 research outputs found

    Weighted ancestors in suffix trees

    Full text link
    The classical, ubiquitous, predecessor problem is to construct a data structure for a set of integers that supports fast predecessor queries. Its generalization to weighted trees, a.k.a. the weighted ancestor problem, has been extensively explored and successfully reduced to the predecessor problem. It is known that any solution for both problems with an input set from a polynomially bounded universe that preprocesses a weighted tree in O(n polylog(n)) space requires \Omega(loglogn) query time. Perhaps the most important and frequent application of the weighted ancestors problem is for suffix trees. It has been a long-standing open question whether the weighted ancestors problem has better bounds for suffix trees. We answer this question positively: we show that a suffix tree built for a text w[1..n] can be preprocessed using O(n) extra space, so that queries can be answered in O(1) time. Thus we improve the running times of several applications. Our improvement is based on a number of data structure tools and a periodicity-based insight into the combinatorial structure of a suffix tree.Comment: 27 pages, LNCS format. A condensed version will appear in ESA 201

    Cross-Document Pattern Matching

    Get PDF
    We study a new variant of the string matching problem called cross-document string matching, which is the problem of indexing a collection of documents to support an efficient search for a pattern in a selected document, where the pattern itself is a substring of another document. Several variants of this problem are considered, and efficient linear-space solutions are proposed with query time bounds that either do not depend at all on the pattern size or depend on it in a very limited way (doubly logarithmic). As a side result, we propose an improved solution to the weighted level ancestor problem

    Efficient LZ78 factorization of grammar compressed text

    Full text link
    We present an efficient algorithm for computing the LZ78 factorization of a text, where the text is represented as a straight line program (SLP), which is a context free grammar in the Chomsky normal form that generates a single string. Given an SLP of size nn representing a text SS of length NN, our algorithm computes the LZ78 factorization of TT in O(nN+mlog⁥N)O(n\sqrt{N}+m\log N) time and O(nN+m)O(n\sqrt{N}+m) space, where mm is the number of resulting LZ78 factors. We also show how to improve the algorithm so that the nNn\sqrt{N} term in the time and space complexities becomes either nLnL, where LL is the length of the longest LZ78 factor, or (N−α)(N - \alpha) where α≄0\alpha \geq 0 is a quantity which depends on the amount of redundancy that the SLP captures with respect to substrings of SS of a certain length. Since m=O(N/logâĄÏƒN)m = O(N/\log_\sigma N) where σ\sigma is the alphabet size, the latter is asymptotically at least as fast as a linear time algorithm which runs on the uncompressed string when σ\sigma is constant, and can be more efficient when the text is compressible, i.e. when mm and nn are small.Comment: SPIRE 201

    Compressed Subsequence Matching and Packed Tree Coloring

    Get PDF
    We present a new algorithm for subsequence matching in grammar compressed strings. Given a grammar of size nn compressing a string of size NN and a pattern string of size mm over an alphabet of size σ\sigma, our algorithm uses O(n+nσw)O(n+\frac{n\sigma}{w}) space and O(n+nσw+mlog⁥Nlog⁥w⋅occ)O(n+\frac{n\sigma}{w}+m\log N\log w\cdot occ) or O(n+nσwlog⁥w+mlog⁥N⋅occ)O(n+\frac{n\sigma}{w}\log w+m\log N\cdot occ) time. Here ww is the word size and occocc is the number of occurrences of the pattern. Our algorithm uses less space than previous algorithms and is also faster for occ=o(nlog⁥N)occ=o(\frac{n}{\log N}) occurrences. The algorithm uses a new data structure that allows us to efficiently find the next occurrence of a given character after a given position in a compressed string. This data structure in turn is based on a new data structure for the tree color problem, where the node colors are packed in bit strings.Comment: To appear at CPM '1

    A Simple Linear-Space Data Structure for Constant-Time Range Minimum Query

    Full text link
    Abstract. We revisit the range minimum query problem and present a new O(n)-space data structure that supports queries in O(1) time. Although previous data structures exist whose asymptotic bounds match ours, our goal is to introduce a new solution that is simple, intuitive, and practical without increasing asymptotic costs for query time or space

    Measurement of ΜˉΌ\bar{\nu}_{\mu} and ΜΌ\nu_{\mu} charged current inclusive cross sections and their ratio with the T2K off-axis near detector

    Get PDF
    We report a measurement of cross section σ(ΜΌ+nucleus→Ό−+X)\sigma(\nu_{\mu}+{\rm nucleus}\rightarrow\mu^{-}+X) and the first measurements of the cross section σ(ΜˉΌ+nucleus→Ό++X)\sigma(\bar{\nu}_{\mu}+{\rm nucleus}\rightarrow\mu^{+}+X) and their ratio R(σ(Μˉ)σ(Îœ))R(\frac{\sigma(\bar \nu)}{\sigma(\nu)}) at (anti-)neutrino energies below 1.5 GeV. We determine the single momentum bin cross section measurements, averaged over the T2K Μˉ/Îœ\bar{\nu}/\nu-flux, for the detector target material (mainly Carbon, Oxygen, Hydrogen and Copper) with phase space restricted laboratory frame kinematics of ΞΌ\theta_{\mu}500 MeV/c. The results are σ(Μˉ)=(0.900±0.029(stat.)±0.088(syst.))×10−39\sigma(\bar{\nu})=\left( 0.900\pm0.029{\rm (stat.)}\pm0.088{\rm (syst.)}\right)\times10^{-39} and $\sigma(\nu)=\left( 2.41\ \pm0.022{\rm{(stat.)}}\pm0.231{\rm (syst.)}\ \right)\times10^{-39}inunitsofcm in units of cm^{2}/nucleonand/nucleon and R\left(\frac{\sigma(\bar{\nu})}{\sigma(\nu)}\right)= 0.373\pm0.012{\rm (stat.)}\pm0.015{\rm (syst.)}$.Comment: 18 pages, 8 figure

    Search for Lorentz and CPT violation using sidereal time dependence of neutrino flavor transitions over a short baseline

    Get PDF
    A class of extensions of the Standard Model allows Lorentz and CPT violations, which can be identified by the observation of sidereal modulations in the neutrino interaction rate. A search for such modulations was performed using the T2K on-axis near detector. Two complementary methods were used in this study, both of which resulted in no evidence of a signal. Limits on associated Lorentz and CPT-violating terms from the Standard Model extension have been derived by taking into account their correlations in this model for the first time. These results imply such symmetry violations are suppressed by a factor of more than 10 20 at the GeV scale

    Volume I. Introduction to DUNE

    Get PDF
    The preponderance of matter over antimatter in the early universe, the dynamics of the supernovae that produced the heavy elements necessary for life, and whether protons eventually decay—these mysteries at the forefront of particle physics and astrophysics are key to understanding the early evolution of our universe, its current state, and its eventual fate. The Deep Underground Neutrino Experiment (DUNE) is an international world-class experiment dedicated to addressing these questions as it searches for leptonic charge-parity symmetry violation, stands ready to capture supernova neutrino bursts, and seeks to observe nucleon decay as a signature of a grand unified theory underlying the standard model. The DUNE far detector technical design report (TDR) describes the DUNE physics program and the technical designs of the single- and dual-phase DUNE liquid argon TPC far detector modules. This TDR is intended to justify the technical choices for the far detector that flow down from the high-level physics goals through requirements at all levels of the Project. Volume I contains an executive summary that introduces the DUNE science program, the far detector and the strategy for its modular designs, and the organization and management of the Project. The remainder of Volume I provides more detail on the science program that drives the choice of detector technologies and on the technologies themselves. It also introduces the designs for the DUNE near detector and the DUNE computing model, for which DUNE is planning design reports. Volume II of this TDR describes DUNE\u27s physics program in detail. Volume III describes the technical coordination required for the far detector design, construction, installation, and integration, and its organizational structure. Volume IV describes the single-phase far detector technology. A planned Volume V will describe the dual-phase technology

    Measurement of the longitudinal diffusion of ionization electrons in the MicroBooNE detector

    Get PDF
    Abstract: Accurate knowledge of electron transport properties is vital to understanding the information provided by liquid argon time projection chambers (LArTPCs). Ionization electron drift-lifetime, local electric field distortions caused by positive ion accumulation, and electron diffusion can all significantly impact the measured signal waveforms. This paper presents a measurement of the effective longitudinal electron diffusion coefficient, DL, in MicroBooNE at the nominal electric field strength of 273.9 V/cm. Historically, this measurement has been made in LArTPC prototype detectors. This represents the first measurement in a large-scale (85 tonne active volume) LArTPC operating in a neutrino beam. This is the largest dataset ever used for this measurement. Using a sample of ∌70,000 through-going cosmic ray muon tracks tagged with MicroBooNE's cosmic ray tagger system, we measure DL = 3.74+0.28 -0.29 cm2/s

    Highly-parallelized simulation of a pixelated LArTPC on a GPU

    Get PDF
    The rapid development of general-purpose computing on graphics processing units (GPGPU) is allowing the implementation of highly-parallelized Monte Carlo simulation chains for particle physics experiments. This technique is particularly suitable for the simulation of a pixelated charge readout for time projection chambers, given the large number of channels that this technology employs. Here we present the first implementation of a full microphysical simulator of a liquid argon time projection chamber (LArTPC) equipped with light readout and pixelated charge readout, developed for the DUNE Near Detector. The software is implemented with an end-to-end set of GPU-optimized algorithms. The algorithms have been written in Python and translated into CUDA kernels using Numba, a just-in-time compiler for a subset of Python and NumPy instructions. The GPU implementation achieves a speed up of four orders of magnitude compared with the equivalent CPU version. The simulation of the current induced on 10^3 pixels takes around 1 ms on the GPU, compared with approximately 10 s on the CPU. The results of the simulation are compared against data from a pixel-readout LArTPC prototype
    • 

    corecore