60,636 research outputs found
The Dynamics of a Genetic Algorithm for a Simple Learning Problem
A formalism for describing the dynamics of Genetic Algorithms (GAs) using
methods from statistical mechanics is applied to the problem of generalization
in a perceptron with binary weights. The dynamics are solved for the case where
a new batch of training patterns is presented to each population member each
generation, which considerably simplifies the calculation. The theory is shown
to agree closely to simulations of a real GA averaged over many runs,
accurately predicting the mean best solution found. For weak selection and
large problem size the difference equations describing the dynamics can be
expressed analytically and we find that the effects of noise due to the finite
size of each training batch can be removed by increasing the population size
appropriately. If this population resizing is used, one can deduce the most
computationally efficient size of training batch each generation. For
independent patterns this choice also gives the minimum total number of
training patterns used. Although using independent patterns is a very
inefficient use of training patterns in general, this work may also prove
useful for determining the optimum batch size in the case where patterns are
recycled.Comment: 28 pages, 4 Postscript figures. Latex using IOP macros ioplppt and
iopl12 which are included. To appear in Journal of Physics A. Also available
at ftp://ftp.cs.man.ac.uk/pub/ai/jls/GAlearn.ps.gz and
http://www.cs.man.ac.uk/~jl
Closed-Loop Statistical Verification of Stochastic Nonlinear Systems Subject to Parametric Uncertainties
This paper proposes a statistical verification framework using Gaussian
processes (GPs) for simulation-based verification of stochastic nonlinear
systems with parametric uncertainties. Given a small number of stochastic
simulations, the proposed framework constructs a GP regression model and
predicts the system's performance over the entire set of possible
uncertainties. Included in the framework is a new metric to estimate the
confidence in those predictions based on the variance of the GP's cumulative
distribution function. This variance-based metric forms the basis of active
sampling algorithms that aim to minimize prediction error through careful
selection of simulations. In three case studies, the new active sampling
algorithms demonstrate up to a 35% improvement in prediction error over other
approaches and are able to correctly identify regions with low prediction
confidence through the variance metric.Comment: 8 pages, submitted to ACC 201
Agile Autonomous Driving using End-to-End Deep Imitation Learning
We present an end-to-end imitation learning system for agile, off-road
autonomous driving using only low-cost sensors. By imitating a model predictive
controller equipped with advanced sensors, we train a deep neural network
control policy to map raw, high-dimensional observations to continuous steering
and throttle commands. Compared with recent approaches to similar tasks, our
method requires neither state estimation nor on-the-fly planning to navigate
the vehicle. Our approach relies on, and experimentally validates, recent
imitation learning theory. Empirically, we show that policies trained with
online imitation learning overcome well-known challenges related to covariate
shift and generalize better than policies trained with batch imitation
learning. Built on these insights, our autonomous driving system demonstrates
successful high-speed off-road driving, matching the state-of-the-art
performance.Comment: 13 pages, Robotics: Science and Systems (RSS) 201
Statistical Mechanics of Dilute Batch Minority Games with Random External Information
We study the dynamics and statics of a dilute batch minority game with random
external information. We focus on the case in which the number of connections
per agent is infinite in the thermodynamic limit. The dynamical scenario of
ergodicity breaking in this model is different from the phase transition in the
standard minority game and is characterised by the onset of long-term memory at
finite integrated response. We demonstrate that finite memory appears at the
AT-line obtained from the corresponding replica calculation, and compare the
behaviour of the dilute model with the minority game with market impact
correction, which is known to exhibit similar features.Comment: 22 pages, 6 figures, text modified, references updated and added,
figure added, typos correcte
- …