1,547 research outputs found

    No Free Lunch versus Occam's Razor in Supervised Learning

    Full text link
    The No Free Lunch theorems are often used to argue that domain specific knowledge is required to design successful algorithms. We use algorithmic information theory to argue the case for a universal bias allowing an algorithm to succeed in all interesting problem domains. Additionally, we give a new algorithm for off-line classification, inspired by Solomonoff induction, with good performance on all structured problems under reasonable assumptions. This includes a proof of the efficacy of the well-known heuristic of randomly selecting training data in the hope of reducing misclassification rates.Comment: 16 LaTeX pages, 1 figur

    Transition asymptotics for reaction-diffusion in random media

    Full text link
    We describe a universal transition mechanism characterizing the passage to an annealed behavior and to a regime where the fluctuations about this behavior are Gaussian, for the long time asymptotics of the empirical average of the expected value of the number of random walks which branch and annihilate on Zd{\mathbb Z}^d, with stationary random rates. The random walks are independent, continuous time rate 2dκ2d\kappa, simple, symmetric, with κ0\kappa \ge 0. A random walk at xZdx\in{\mathbb Z}^d, binary branches at rate v+(x)v_+(x), and annihilates at rate v(x)v_-(x). The random environment ww has coordinates w(x)=(v(x),v+(x))w(x)=(v_-(x),v_+(x)) which are i.i.d. We identify a natural way to describe the annealed-Gaussian transition mechanism under mild conditions on the rates. Indeed, we introduce the exponents Fθ(t):=H1((1+θ)t)(1+θ)H1(t)θF_\theta(t):=\frac{H_1((1+\theta)t)-(1+\theta)H_1(t)}{\theta}, and assume that F2θ(t)Fθ(t)θlog(κt+e)\frac{F_{2\theta}(t)-F_\theta(t)}{\theta\log(\kappa t+e)}\to\infty for θ>0|\theta|>0 small enough, where H1(t):=logH_1(t):=\log and denotes the average of the expected value of the number of particles m(0,t,w)m(0,t,w) at time tt and an environment of rates ww, given that initially there was only one particle at 0. Then the empirical average of m(x,t,w)m(x,t,w) over a box of side L(t)L(t) has different behaviors: if L(t)e1dFϵ(t) L(t)\ge e^{\frac{1}{d} F_\epsilon(t)} for some ϵ>0\epsilon >0 and large enough tt, a law of large numbers is satisfied; if L(t)e1dFϵ(2t) L(t)\ge e^{\frac{1}{d} F_\epsilon (2t)} for some ϵ>0\epsilon>0 and large enough tt, a CLT is satisfied. These statements are violated if the reversed inequalities are satisfied for some negative ϵ\epsilon. Applications to potentials with Weibull, Frechet and double exponential tails are given.Comment: To appear in: Probability and Mathematical Physics: A Volume in Honor of Stanislav Molchanov, Editors - AMS | CRM, (2007
    corecore