144 research outputs found
Improved Approximate String Matching and Regular Expression Matching on Ziv-Lempel Compressed Texts
We study the approximate string matching and regular expression matching
problem for the case when the text to be searched is compressed with the
Ziv-Lempel adaptive dictionary compression schemes. We present a time-space
trade-off that leads to algorithms improving the previously known complexities
for both problems. In particular, we significantly improve the space bounds,
which in practical applications are likely to be a bottleneck
Fast Searching in Packed Strings
Given strings and the (exact) string matching problem is to find all
positions of substrings in matching . The classical Knuth-Morris-Pratt
algorithm [SIAM J. Comput., 1977] solves the string matching problem in linear
time which is optimal if we can only read one character at the time. However,
most strings are stored in a computer in a packed representation with several
characters in a single word, giving us the opportunity to read multiple
characters simultaneously. In this paper we study the worst-case complexity of
string matching on strings given in packed representation. Let be
the lengths and , respectively, and let denote the size of the
alphabet. On a standard unit-cost word-RAM with logarithmic word size we
present an algorithm using time O\left(\frac{n}{\log_\sigma n} + m +
\occ\right). Here \occ is the number of occurrences of in . For this improves the bound of the Knuth-Morris-Pratt algorithm.
Furthermore, if our algorithm is optimal since any
algorithm must spend at least \Omega(\frac{(n+m)\log
\sigma}{\log n} + \occ) = \Omega(\frac{n}{\log_\sigma n} + \occ) time to
read the input and report all occurrences. The result is obtained by a novel
automaton construction based on the Knuth-Morris-Pratt algorithm combined with
a new compact representation of subautomata allowing an optimal
tabulation-based simulation.Comment: To appear in Journal of Discrete Algorithms. Special Issue on CPM
200
Measurement of the Transverse Spin Transfer Coefficient D_nn(0˚) for (p,n) Reactions at 160 MeV
This research was sponsored by the National Science Foundation Grant NSF PHY 87-1440
Spin Transfer Measurements for (p,n) Reactions at Intermediate Energy
This research was sponsored by the National Science Foundation Grant NSF PHY 87-1440
Developing a dementia-specific preference-based quality of life measure (AD-5D) in Australia: a valuation study protocol
Introduction: Generic instruments for assessing health-related quality of life may lack the sensitivity to detect changes in health specific to certain conditions, such as dementia. The QOL-AD is a widely used and well validated condition-specific instrument for assessing health-related quality of life for people
living with dementia, but it does not enable the calculation of Quality Adjusted Life Years (QALYs), the
basis of cost utility analysis. This study will generate a preference-based scoring algorithm for a health
state classification system (the AD-5D) derived from the QOL-AD.
Methods and analysis: Discrete choice experiments with duration (DCETTO) and best-worst scaling
(BWS) health state valuation tasks will be administered to a representative sample of 2,000 members of
the Australian general population via an online survey and to 250 dementia dyads (250 people with
dementia and their carers) via face-to-face interview. A multinomial (conditional) logistic framework will
be used to analyse responses and produce the utility algorithm for the AD-5D.
Ethics and dissemination: The algorithms developed will enable prospective and retrospective
economic evaluation of any treatment or intervention targeting people with dementia where the QOL-AD
has been administered and will be available online. Results will be disseminated through journals that
publish health economics articles and through professional conferences. The study has ethical approval
Measurements of Gamow-Teller Strength Distributions in Masses 13 and 15
This research was sponsored by the National Science Foundation Grant NSF PHY 87-1440
A molecular dynamics simulation of polymer crystallization from oriented amorphous state
Molecular process of crystallization from an oriented amorphous state was
reproduced by molecular dynamics simulation for a realistic polyethylene model.
Initial oriented amorphous state was obtained by uniaxial drawing an isotropic
glassy state at 100 K. By the temperature jump from 100 K to 330 K, there
occurred the crystallization into the fiber structure, during the process of
which we observed the developments of various order parameters. The real space
image and its Fourier transform revealed that a hexagonally ordered domain was
initially formed, and then highly ordered crystalline state with stacked
lamellae developed after further adjustment of the relative heights of the
chains along their axes.Comment: 4 pages, 3 figure
Mixing Bandt-Pompe and Lempel-Ziv approaches: another way to analyze the complexity of continuous-states sequences
In this paper, we propose to mix the approach underlying Bandt-Pompe
permutation entropy with Lempel-Ziv complexity, to design what we call
Lempel-Ziv permutation complexity. The principle consists of two steps: (i)
transformation of a continuous-state series that is intrinsically multivariate
or arises from embedding into a sequence of permutation vectors, where the
components are the positions of the components of the initial vector when
re-arranged; (ii) performing the Lempel-Ziv complexity for this series of
`symbols', as part of a discrete finite-size alphabet. On the one hand, the
permutation entropy of Bandt-Pompe aims at the study of the entropy of such a
sequence; i.e., the entropy of patterns in a sequence (e.g., local increases or
decreases). On the other hand, the Lempel-Ziv complexity of a discrete-state
sequence aims at the study of the temporal organization of the symbols (i.e.,
the rate of compressibility of the sequence). Thus, the Lempel-Ziv permutation
complexity aims to take advantage of both of these methods. The potential from
such a combined approach - of a permutation procedure and a complexity analysis
- is evaluated through the illustration of some simulated data and some real
data. In both cases, we compare the individual approaches and the combined
approach.Comment: 30 pages, 4 figure
LAGraph: Linear algebra, network analysis libraries, and the study of graph algorithms
Graph algorithms can be expressed in terms of linear algebra. GraphBLAS is a library of low-level building blocks for such algorithms that targets algorithm developers. LAGraph builds on top of the GraphBLAS to target users of graph algorithms with high-level algorithms common in network analysis. In this paper, we describe the first release of the LAGraph library, the design decisions behind the library, and performance using the GAP benchmark suite. LAGraph, however, is much more than a library. It is also a project to document and analyze the full range of algorithms enabled by the GraphBLAS. To that end, we have developed a compact and intuitive notation for describing these algorithms. In this paper, we present that notation with examples from the GAP benchmark suite
LAGraph: Linear algebra, network analysis libraries, and the study of graph algorithms
Graph algorithms can be expressed in terms of linear algebra. GraphBLAS is a library of low-level building blocks for such algorithms that targets algorithm developers. LAGraph builds on top of the GraphBLAS to target users of graph algorithms with high-level algorithms common in network analysis. In this paper, we describe the first release of the LAGraph library, the design decisions behind the library, and performance using the GAP benchmark suite. LAGraph, however, is much more than a library. It is also a project to document and analyze the full range of algorithms enabled by the GraphBLAS. To that end, we have developed a compact and intuitive notation for describing these algorithms. In this paper, we present that notation with examples from the GAP benchmark suite
- …