7,104 research outputs found
Reinforcement Learning using Augmented Neural Networks
Neural networks allow Q-learning reinforcement learning agents such as deep
Q-networks (DQN) to approximate complex mappings from state spaces to value
functions. However, this also brings drawbacks when compared to other function
approximators such as tile coding or their generalisations, radial basis
functions (RBF) because they introduce instability due to the side effect of
globalised updates present in neural networks. This instability does not even
vanish in neural networks that do not have any hidden layers. In this paper, we
show that simple modifications to the structure of the neural network can
improve stability of DQN learning when a multi-layer perceptron is used for
function approximation.Comment: 7 pages; two columns; 4 figure
A survey of outlier detection methodologies
Outlier detection has been used for centuries to detect and, where appropriate, remove anomalous observations from data. Outliers arise due to mechanical faults, changes in system behaviour, fraudulent behaviour, human error, instrument error or simply through natural deviations in populations. Their detection can identify system faults and fraud before they escalate with potentially catastrophic consequences. It can identify errors and remove their contaminating effect on the data set and as such to purify the data for processing. The original outlier detection methods were arbitrary but now, principled and systematic techniques are used, drawn from the full gamut of Computer Science and Statistics. In this paper, we introduce a survey of contemporary techniques for outlier detection. We identify their respective motivations and distinguish their advantages and disadvantages in a comparative review
Info Navigator: A visualization tool for document searching and browsing
In this paper we investigate the retrieval performance of monophonic and polyphonic queries made on a polyphonic music database. We extend the n-gram approach for full-music indexing of monophonic music data to polyphonic music using both rhythm and pitch information. We define an experimental framework for a comparative and fault-tolerance study of various n-gramming strategies and encoding levels. For monophonic queries, we focus in particular on query-by-humming systems, and for polyphonic queries on query-by-example. Error models addressed in several studies are surveyed for the fault-tolerance study. Our experiments show that different n-gramming strategies and encoding precision differ widely in their effectiveness. We present the results of our study on a collection of 6366 polyphonic MIDI-encoded music pieces
Dimensional Affect and Expression in Natural and Mediated Interaction
There is a perceived controversy as to whether the cognitive representation
of affect is better modelled using a dimensional or categorical theory. This
paper first suggests that these views are, in fact, compatible. The paper then
discusses this theme and related issues in reference to a commonly stated
application domain of research on human affect and expression: human computer
interaction (HCI). The novel suggestion here is that a more realistic framing
of studies of human affect in expression with reference to HCI and,
particularly HCHI (Human-Computer-Human Interaction) entails some
re-formulation of the approach to the basic phenomena themselves. This theme is
illustrated with several examples from several recent research projects.Comment: Invited article presented at the 23rd Annual Meeting of the
International Society for Psychophysics, Tokyo, Japan, 20-23 October, 2007,
Proceedings of Fechner Day vol. 23 (2007
A supervised clustering approach for fMRI-based inference of brain states
We propose a method that combines signals from many brain regions observed in
functional Magnetic Resonance Imaging (fMRI) to predict the subject's behavior
during a scanning session. Such predictions suffer from the huge number of
brain regions sampled on the voxel grid of standard fMRI data sets: the curse
of dimensionality. Dimensionality reduction is thus needed, but it is often
performed using a univariate feature selection procedure, that handles neither
the spatial structure of the images, nor the multivariate nature of the signal.
By introducing a hierarchical clustering of the brain volume that incorporates
connectivity constraints, we reduce the span of the possible spatial
configurations to a single tree of nested regions tailored to the signal. We
then prune the tree in a supervised setting, hence the name supervised
clustering, in order to extract a parcellation (division of the volume) such
that parcel-based signal averages best predict the target information.
Dimensionality reduction is thus achieved by feature agglomeration, and the
constructed features now provide a multi-scale representation of the signal.
Comparisons with reference methods on both simulated and real data show that
our approach yields higher prediction accuracy than standard voxel-based
approaches. Moreover, the method infers an explicit weighting of the regions
involved in the regression or classification task
Improving accuracy and power with transfer learning using a meta-analytic database
Typical cohorts in brain imaging studies are not large enough for systematic
testing of all the information contained in the images. To build testable
working hypotheses, investigators thus rely on analysis of previous work,
sometimes formalized in a so-called meta-analysis. In brain imaging, this
approach underlies the specification of regions of interest (ROIs) that are
usually selected on the basis of the coordinates of previously detected
effects. In this paper, we propose to use a database of images, rather than
coordinates, and frame the problem as transfer learning: learning a
discriminant model on a reference task to apply it to a different but related
new task. To facilitate statistical analysis of small cohorts, we use a sparse
discriminant model that selects predictive voxels on the reference task and
thus provides a principled procedure to define ROIs. The benefits of our
approach are twofold. First it uses the reference database for prediction, i.e.
to provide potential biomarkers in a clinical setting. Second it increases
statistical power on the new task. We demonstrate on a set of 18 pairs of
functional MRI experimental conditions that our approach gives good prediction.
In addition, on a specific transfer situation involving different scanners at
different locations, we show that voxel selection based on transfer learning
leads to higher detection power on small cohorts.Comment: MICCAI, Nice : France (2012
Cleaning Up the Kitchen Sink: Growth Empirics When the World Is Not Simple
This paper explores the relevance of unknown nonlinearities for growth empirics. Recent theoretical contributions and case-study evidence suggest that nonlinearities are pervasive in the growth process. I show that the postwar data provide strong evidence in favor of generalized non-linearities. I provide two alternative mechanisms for making inference about the effects of production-function shifters on growth that do not make a priori assumptions about functional form: monotonicity tests and average derivative estimation. The results of these tests point towards a greater role for structural variables and a smaller role for policy variables than the linear model.Economic Growth, Cross-Country Growth Regressions, Non-linearities, Non-parametric econometrics
- …