47,471 research outputs found
Machine learning approach for quantum non-Markovian noise classification
In this paper, machine learning and artificial neural network models are
proposed for quantum noise classification in stochastic quantum dynamics. For
this purpose, we train and then validate support vector machine, multi-layer
perceptron and recurrent neural network, models with different complexity and
accuracy, to solve supervised binary classification problems. By exploiting the
quantum random walk formalism, we demonstrate the high efficacy of such tools
in classifying noisy quantum dynamics using data sets collected in a single
realisation of the quantum system evolution. In addition, we also show that for
a successful classification one just needs to measure, in a sequence of
discrete time instants, the probabilities that the analysed quantum system is
in one of the allowed positions or energy configurations, without any external
driving. Thus, neither measurements of quantum coherences nor sequences of
control pulses are required. Since in principle the training of the machine
learning models can be performed a-priori on synthetic data, our approach is
expected to find direct application in a vast number of experimental schemes
and also for the noise benchmarking of the already available noisy
intermediate-scale quantum devices.Comment: 14 pages, 3 figures, 3 table
Machine learning for scientific data mining and solar eruption prediction
This dissertation explores new machine learning techniques and adapts them to mine scientific data, specifically data from solar physics and space weather studies. The dissertation tackles three important problems in heliophysics: solar flare prediction, coronal mass ejection (CME) prediction and Stokes inversion.
First, the dissertation presents a long short-term memory (LSTM) network for predicting whether an active region (AR) would produce a certain class of solar flare within the next 24 hours. The essence of this approach is to model data samples in an AR as time series and use LSTMs to capture temporal information of the data samples. The LSTM network consists of an LSTM layer, an attention layer, two fully connected layers and an output layer. The attention layer is designed to allow the LSTM network to automatically search for parts of the data samples that are related to the prediction of solar flares.
Second, the dissertation presents two recurrent neural networks (RNNs), one based on gated recurrent units and the other based on LSTM, for predicting whether an AR that produces a significant flare will also initiate a CME. Again, data samples in an AR are modeled as time series and the RNNs are used to capture temporal dependencies in the time series. A feature selection technique is employed to enhance prediction accuracy.
Third, the dissertation approaches the Stokes inversion problem using a novel convolutional neural network (CNN). This CNN method is faster, and produces cleaner magnetic maps, than a widely used physics-based tool. Furthermore, the CNN method outperforms other machine learning algorithms such as multiple support vector regression and multilayer perceptrons.
Findings reported here have been validated by substantial experiments based on different datasets. The dissertation concludes with a fully operational database system containing real-time flare forecasting results produced by the proposed LSTM method. This is the first cyberinfrastructure capable of continuous learning and forecasting of solar flares based on deep learning
Classification of Occluded Objects using Fast Recurrent Processing
Recurrent neural networks are powerful tools for handling incomplete data
problems in computer vision, thanks to their significant generative
capabilities. However, the computational demand for these algorithms is too
high to work in real time, without specialized hardware or software solutions.
In this paper, we propose a framework for augmenting recurrent processing
capabilities into a feedforward network without sacrificing much from
computational efficiency. We assume a mixture model and generate samples of the
last hidden layer according to the class decisions of the output layer, modify
the hidden layer activity using the samples, and propagate to lower layers. For
visual occlusion problem, the iterative procedure emulates feedforward-feedback
loop, filling-in the missing hidden layer activity with meaningful
representations. The proposed algorithm is tested on a widely used dataset, and
shown to achieve 2 improvement in classification accuracy for occluded
objects. When compared to Restricted Boltzmann Machines, our algorithm shows
superior performance for occluded object classification.Comment: arXiv admin note: text overlap with arXiv:1409.8576 by other author
Modeling Financial Time Series with Artificial Neural Networks
Financial time series convey the decisions and actions of a population of human actors over time. Econometric and regressive models have been developed in the past decades for analyzing these time series. More recently, biologically inspired artificial neural network models have been shown to overcome some of the main challenges of traditional techniques by better exploiting the non-linear, non-stationary, and oscillatory nature of noisy, chaotic human interactions. This review paper explores the options, benefits, and weaknesses of the various forms of artificial neural networks as compared with regression techniques in the field of financial time series analysis.CELEST, a National Science Foundation Science of Learning Center (SBE-0354378); SyNAPSE program of the Defense Advanced Research Project Agency (HR001109-03-0001
Conversion of Artificial Recurrent Neural Networks to Spiking Neural Networks for Low-power Neuromorphic Hardware
In recent years the field of neuromorphic low-power systems that consume
orders of magnitude less power gained significant momentum. However, their
wider use is still hindered by the lack of algorithms that can harness the
strengths of such architectures. While neuromorphic adaptations of
representation learning algorithms are now emerging, efficient processing of
temporal sequences or variable length-inputs remain difficult. Recurrent neural
networks (RNN) are widely used in machine learning to solve a variety of
sequence learning tasks. In this work we present a train-and-constrain
methodology that enables the mapping of machine learned (Elman) RNNs on a
substrate of spiking neurons, while being compatible with the capabilities of
current and near-future neuromorphic systems. This "train-and-constrain" method
consists of first training RNNs using backpropagation through time, then
discretizing the weights and finally converting them to spiking RNNs by
matching the responses of artificial neurons with those of the spiking neurons.
We demonstrate our approach by mapping a natural language processing task
(question classification), where we demonstrate the entire mapping process of
the recurrent layer of the network on IBM's Neurosynaptic System "TrueNorth", a
spike-based digital neuromorphic hardware architecture. TrueNorth imposes
specific constraints on connectivity, neural and synaptic parameters. To
satisfy these constraints, it was necessary to discretize the synaptic weights
and neural activities to 16 levels, and to limit fan-in to 64 inputs. We find
that short synaptic delays are sufficient to implement the dynamical (temporal)
aspect of the RNN in the question classification task. The hardware-constrained
model achieved 74% accuracy in question classification while using less than
0.025% of the cores on one TrueNorth chip, resulting in an estimated power
consumption of ~17 uW
- …