542 research outputs found
Long and short paths in uniform random recursive dags
In a uniform random recursive k-dag, there is a root, 0, and each node in
turn, from 1 to n, chooses k uniform random parents from among the nodes of
smaller index. If S_n is the shortest path distance from node n to the root,
then we determine the constant \sigma such that S_n/log(n) tends to \sigma in
probability as n tends to infinity. We also show that max_{1 \le i \le n}
S_i/log(n) tends to \sigma in probability.Comment: 16 page
Traveling Waves, Front Selection, and Exact Nontrivial Exponents in a Random Fragmentation Problem
We study a random bisection problem where an initial interval of length x is
cut into two random fragments at the first stage, then each of these two
fragments is cut further, etc. We compute the probability P_n(x) that at the
n-th stage, each of the 2^n fragments is shorter than 1. We show that P_n(x)
approaches a traveling wave form, and the front position x_n increases as
x_n\sim n^{\beta}{\rho}^n for large n. We compute exactly the exponents
\rho=1.261076... and \beta=0.453025.... as roots of transcendental equations.
We also solve the m-section problem where each interval is broken into m
fragments. In particular, the generalized exponents grow as \rho_m\approx
m/(\ln m) and \beta_m\approx 3/(2\ln m) in the large m limit. Our approach
establishes an intriguing connection between extreme value statistics and
traveling wave propagation in the context of the fragmentation problem.Comment: 4 pages Revte
PAC-Bayesian Bounds for Randomized Empirical Risk Minimizers
The aim of this paper is to generalize the PAC-Bayesian theorems proved by
Catoni in the classification setting to more general problems of statistical
inference. We show how to control the deviations of the risk of randomized
estimators. A particular attention is paid to randomized estimators drawn in a
small neighborhood of classical estimators, whose study leads to control the
risk of the latter. These results allow to bound the risk of very general
estimation procedures, as well as to perform model selection
Non-parametric comparison of histogrammed two-dimensional data distributions using the Energy Test
When monitoring complex experiments, comparison is often made between regularly acquired histograms of data and reference histograms which represent the ideal state of the equipment. With the larger HEP experiments now ramping up, there is a need for automation of this task since the volume of comparisons could overwhelm human operators. However, the two-dimensional histogram comparison tools available in ROOT have been noted in the past to exhibit shortcomings. We discuss a newer comparison test for two-dimensional histograms, based on the Energy Test of Aslan and Zech, which provides more conclusive
discrimination between histograms of data coming from different distributions than methods provided in a recent ROOT release.The Science and Technology Facilities Council, U
A Fast Algorithm Finding the Shortest Reset Words
In this paper we present a new fast algorithm finding minimal reset words for
finite synchronizing automata. The problem is know to be computationally hard,
and our algorithm is exponential. Yet, it is faster than the algorithms used so
far and it works well in practice. The main idea is to use a bidirectional BFS
and radix (Patricia) tries to store and compare resulted subsets. We give both
theoretical and practical arguments showing that the branching factor is
reduced efficiently. As a practical test we perform an experimental study of
the length of the shortest reset word for random automata with states and 2
input letters. We follow Skvorsov and Tipikin, who have performed such a study
using a SAT solver and considering automata up to states. With our
algorithm we are able to consider much larger sample of automata with up to
states. In particular, we obtain a new more precise estimation of the
expected length of the shortest reset word .Comment: COCOON 2013. The final publication is available at
http://link.springer.com/chapter/10.1007%2F978-3-642-38768-5_1
Stationary probability density of stochastic search processes in global optimization
A method for the construction of approximate analytical expressions for the
stationary marginal densities of general stochastic search processes is
proposed. By the marginal densities, regions of the search space that with high
probability contain the global optima can be readily defined. The density
estimation procedure involves a controlled number of linear operations, with a
computational cost per iteration that grows linearly with problem size
A novel approach to light-front perturbation theory
We suggest a possible algorithm to calculate one-loop n-point functions
within a variant of light-front perturbation theory. The key ingredients are
the covariant Passarino-Veltman scheme and a surprising integration formula
that localises Feynman integrals at vanishing longitudinal momentum. The
resulting expressions are generalisations of Weinberg's infinite-momentum
results and are manifestly Lorentz invariant. For n = 2 and 3 we explicitly
show how to relate those to light-front integrals with standard energy
denominators. All expressions are rendered finite by means of transverse
dimensional regularisation.Comment: 10 pages, 5 figure
Simulation of truncated normal variables
We provide in this paper simulation algorithms for one-sided and two-sided
truncated normal distributions. These algorithms are then used to simulate
multivariate normal variables with restricted parameter space for any
covariance structure.Comment: This 1992 paper appeared in 1995 in Statistics and Computing and the
gist of it is contained in Monte Carlo Statistical Methods (2004), but I
receive weekly requests for reprints so here it is
Wear Minimization for Cuckoo Hashing: How Not to Throw a Lot of Eggs into One Basket
We study wear-leveling techniques for cuckoo hashing, showing that it is
possible to achieve a memory wear bound of after the
insertion of items into a table of size for a suitable constant
using cuckoo hashing. Moreover, we study our cuckoo hashing method empirically,
showing that it significantly improves on the memory wear performance for
classic cuckoo hashing and linear probing in practice.Comment: 13 pages, 1 table, 7 figures; to appear at the 13th Symposium on
Experimental Algorithms (SEA 2014
On the flexibility of the design of Multiple Try Metropolis schemes
The Multiple Try Metropolis (MTM) method is a generalization of the classical
Metropolis-Hastings algorithm in which the next state of the chain is chosen
among a set of samples, according to normalized weights. In the literature,
several extensions have been proposed. In this work, we show and remark upon
the flexibility of the design of MTM-type methods, fulfilling the detailed
balance condition. We discuss several possibilities and show different
numerical results
- âŠ