241,602 research outputs found
Mean-payoff Automaton Expressions
Quantitative languages are an extension of boolean languages that assign to
each word a real number. Mean-payoff automata are finite automata with
numerical weights on transitions that assign to each infinite path the long-run
average of the transition weights. When the mode of branching of the automaton
is deterministic, nondeterministic, or alternating, the corresponding class of
quantitative languages is not robust as it is not closed under the pointwise
operations of max, min, sum, and numerical complement. Nondeterministic and
alternating mean-payoff automata are not decidable either, as the quantitative
generalization of the problems of universality and language inclusion is
undecidable.
We introduce a new class of quantitative languages, defined by mean-payoff
automaton expressions, which is robust and decidable: it is closed under the
four pointwise operations, and we show that all decision problems are decidable
for this class. Mean-payoff automaton expressions subsume deterministic
mean-payoff automata, and we show that they have expressive power incomparable
to nondeterministic and alternating mean-payoff automata. We also present for
the first time an algorithm to compute distance between two quantitative
languages, and in our case the quantitative languages are given as mean-payoff
automaton expressions
SlowFuzz: Automated Domain-Independent Detection of Algorithmic Complexity Vulnerabilities
Algorithmic complexity vulnerabilities occur when the worst-case time/space
complexity of an application is significantly higher than the respective
average case for particular user-controlled inputs. When such conditions are
met, an attacker can launch Denial-of-Service attacks against a vulnerable
application by providing inputs that trigger the worst-case behavior. Such
attacks have been known to have serious effects on production systems, take
down entire websites, or lead to bypasses of Web Application Firewalls.
Unfortunately, existing detection mechanisms for algorithmic complexity
vulnerabilities are domain-specific and often require significant manual
effort. In this paper, we design, implement, and evaluate SlowFuzz, a
domain-independent framework for automatically finding algorithmic complexity
vulnerabilities. SlowFuzz automatically finds inputs that trigger worst-case
algorithmic behavior in the tested binary. SlowFuzz uses resource-usage-guided
evolutionary search techniques to automatically find inputs that maximize
computational resource utilization for a given application.Comment: ACM CCS '17, October 30-November 3, 2017, Dallas, TX, US
An automatic tool flow for the combined implementation of multi-mode circuits
A multi-mode circuit implements the functionality of a limited number of circuits, called modes, of which at any given time only one needs to be realised. Using run-time reconfiguration of an FPGA, all the modes can be implemented on the same reconfigurable region, requiring only an area that can contain the biggest mode. Typically, conventional run-time reconfiguration techniques generate a configuration for every mode separately. To switch between modes the complete reconfigurable region is rewritten, which often leads to very long reconfiguration times. In this paper we present a novel, fully automated tool flow that exploits similarities between the modes and uses Dynamic Circuit Specialization to drastically reduce reconfiguration time. Experimental results show that the number of bits that is rewritten in the configuration memory reduces with a factor from 4.6X to 5.1X without significant performance penalties
Field theoretic approach to metastability in the contact process
A quantum field theoretic formulation of the dynamics of the Contact Process
on a regular graph of degree z is introduced. A perturbative calculation in
powers of 1/z of the effective potential for the density of particles phi(t)
and an instantonic field psi(t) emerging from the quantum formalism is
performed. Corrections to the mean-field distribution of densities of particles
in the out-of-equilibrium stationary state are derived in powers of 1/z.
Results for typical (e.g. average density) and rare fluctuation (e.g. lifetime
of the metastable state) properties are in very good agreement with numerical
simulations carried out on D-dimensional hypercubic (z=2D) and Cayley lattices.Comment: Final published version; 20 pages, 5 figure
Recommended from our members
Protected Health Information filter (Philter): accurately and securely de-identifying free-text clinical notes.
There is a great and growing need to ascertain what exactly is the state of a patient, in terms of disease progression, actual care practices, pathology, adverse events, and much more, beyond the paucity of data available in structured medical record data. Ascertaining these harder-to-reach data elements is now critical for the accurate phenotyping of complex traits, detection of adverse outcomes, efficacy of off-label drug use, and longitudinal patient surveillance. Clinical notes often contain the most detailed and relevant digital information about individual patients, the nuances of their diseases, the treatment strategies selected by physicians, and the resulting outcomes. However, notes remain largely unused for research because they contain Protected Health Information (PHI), which is synonymous with individually identifying data. Previous clinical note de-identification approaches have been rigid and still too inaccurate to see any substantial real-world use, primarily because they have been trained with too small medical text corpora. To build a new de-identification tool, we created the largest manually annotated clinical note corpus for PHI and develop a customizable open-source de-identification software called Philter ("Protected Health Information filter"). Here we describe the design and evaluation of Philter, and show how it offers substantial real-world improvements over prior methods
- …