3,599 research outputs found
Recommended from our members
Why a Single Parallelization Strategy Is Not Enough in Knowledge Bases
We address the problem of parallelizing the evaluation of logic programs in data intensive applications. We argue that the appropriate parallelization strategy for logic-program evaluation depends on the program being evaluated. Therefore, this paper is concerned with the issues of program classification and parallelization strategies. We propose several parallelization strategies based on the concept of data reduction—the original logic program is evaluated by several processors working in parallel, each using only a subset of the database. The strategies differ on the evaluation cost, the overhead of communication and synchronization among processors, and the programs to which they are applicable. In particular, we start our study with pure parallelization, i.e., parallelization without overhead. An interesting class structure of logic programs is demonstrated, when considering amenability to pure parallelization. The relationship to the NC complexity class is demonstrated. Then we propose strategies that do incur an overhead, but are optimal in a sense that will be precisely defined. This paper makes the initial steps towards a theory of parallel logic programming
Computing the eigenvalues of symmetric H2-matrices by slicing the spectrum
The computation of eigenvalues of large-scale matrices arising from finite
element discretizations has gained significant interest in the last decade.
Here we present a new algorithm based on slicing the spectrum that takes
advantage of the rank structure of resolvent matrices in order to compute m
eigenvalues of the generalized symmetric eigenvalue problem in operations, where is a small constant
Query Rewriting and Optimization for Ontological Databases
Ontological queries are evaluated against a knowledge base consisting of an
extensional database and an ontology (i.e., a set of logical assertions and
constraints which derive new intensional knowledge from the extensional
database), rather than directly on the extensional database. The evaluation and
optimization of such queries is an intriguing new problem for database
research. In this paper, we discuss two important aspects of this problem:
query rewriting and query optimization. Query rewriting consists of the
compilation of an ontological query into an equivalent first-order query
against the underlying extensional database. We present a novel query rewriting
algorithm for rather general types of ontological constraints which is
well-suited for practical implementations. In particular, we show how a
conjunctive query against a knowledge base, expressed using linear and sticky
existential rules, that is, members of the recently introduced Datalog+/-
family of ontology languages, can be compiled into a union of conjunctive
queries (UCQ) against the underlying database. Ontological query optimization,
in this context, attempts to improve this rewriting process so to produce
possibly small and cost-effective UCQ rewritings for an input query.Comment: arXiv admin note: text overlap with arXiv:1312.5914 by other author
On the Security of Y-00 under Fast Correlation and Other Attacks on the Key
The potential weakness of the Y-00 direct encryption protocol when the
encryption box ENC in Y-00 is not chosen properly is demonstrated in a fast
correlation attack by S. Donnet et al in Phys. Lett. A 35, 6 (2006) 406-410. In
this paper, we show how this weakness can be eliminated with a proper design of
ENC. In particular, we present a Y-00 configuration that is more secure than
AES under known-plaintext attack. It is also shown that under any
ciphertext-only attack, full information-theoretic security on the Y-00 seed
key is obtained for any ENC when proper deliberate signal randomization is
employed
Space--Time Tradeoffs for Subset Sum: An Improved Worst Case Algorithm
The technique of Schroeppel and Shamir (SICOMP, 1981) has long been the most
efficient way to trade space against time for the SUBSET SUM problem. In the
random-instance setting, however, improved tradeoffs exist. In particular, the
recently discovered dissection method of Dinur et al. (CRYPTO 2012) yields a
significantly improved space--time tradeoff curve for instances with strong
randomness properties. Our main result is that these strong randomness
assumptions can be removed, obtaining the same space--time tradeoffs in the
worst case. We also show that for small space usage the dissection algorithm
can be almost fully parallelized. Our strategy for dealing with arbitrary
instances is to instead inject the randomness into the dissection process
itself by working over a carefully selected but random composite modulus, and
to introduce explicit space--time controls into the algorithm by means of a
"bailout mechanism"
- …