1,874 research outputs found
Enhancing Stock Movement Prediction with Adversarial Training
This paper contributes a new machine learning solution for stock movement
prediction, which aims to predict whether the price of a stock will be up or
down in the near future. The key novelty is that we propose to employ
adversarial training to improve the generalization of a neural network
prediction model. The rationality of adversarial training here is that the
input features to stock prediction are typically based on stock price, which is
essentially a stochastic variable and continuously changed with time by nature.
As such, normal training with static price-based features (e.g. the close
price) can easily overfit the data, being insufficient to obtain reliable
models. To address this problem, we propose to add perturbations to simulate
the stochasticity of price variable, and train the model to work well under
small yet intentional perturbations. Extensive experiments on two real-world
stock data show that our method outperforms the state-of-the-art solution with
3.11% relative improvements on average w.r.t. accuracy, validating the
usefulness of adversarial training for stock prediction task.Comment: IJCAI 201
A Novel Distributed Representation of News (DRNews) for Stock Market Predictions
In this study, a novel Distributed Representation of News (DRNews) model is
developed and applied in deep learning-based stock market predictions. With the
merit of integrating contextual information and cross-documental knowledge, the
DRNews model creates news vectors that describe both the semantic information
and potential linkages among news events through an attributed news network.
Two stock market prediction tasks, namely the short-term stock movement
prediction and stock crises early warning, are implemented in the framework of
the attention-based Long Short Term-Memory (LSTM) network. It is suggested that
DRNews substantially enhances the results of both tasks comparing with five
baselines of news embedding models. Further, the attention mechanism suggests
that short-term stock trend and stock market crises both receive influences
from daily news with the former demonstrates more critical responses on the
information related to the stock market {\em per se}, whilst the latter draws
more concerns on the banking sector and economic policies.Comment: 25 page
Enhanced news sentiment analysis using deep learning methods
We explore the predictive power of historical news sentiments based on financial market performance to forecast financial news sentiments. We define news sentiments based on stock price returns averaged over one minute right after a news article has been released. If the stock price exhibits positive (negative) return, we classify the news article released just prior to the observed stock return as positive (negative). We use Wikipedia and Gigaword five corpus articles from 2014 and we apply the global vectors for word representation method to this corpus to create word vectors to use as inputs into the deep learning TensorFlow network. We analyze high-frequency (intraday) Thompson Reuters News Archive as well as the high-frequency price tick history of the Dow Jones Industrial Average (DJIA 30) Index individual stocks for the period between 1/1/2003 and 12/30/2013. We apply a combination of deep learning methodologies of recurrent neural network with long short-term memory units to train the Thompson Reuters News Archive Data from 2003 to 2012, and we test the forecasting power of our method on 2013 News Archive data. We find that the forecasting accuracy of our methodology improves when we switch from random selection of positive and negative news to selecting the news with highest positive scores as positive news and news with highest negative scores as negative news to create our training data set.Published versio
Exploring Interpretable LSTM Neural Networks over Multi-Variable Data
For recurrent neural networks trained on time series with target and
exogenous variables, in addition to accurate prediction, it is also desired to
provide interpretable insights into the data. In this paper, we explore the
structure of LSTM recurrent neural networks to learn variable-wise hidden
states, with the aim to capture different dynamics in multi-variable time
series and distinguish the contribution of variables to the prediction. With
these variable-wise hidden states, a mixture attention mechanism is proposed to
model the generative process of the target. Then we develop associated training
methods to jointly learn network parameters, variable and temporal importance
w.r.t the prediction of the target variable. Extensive experiments on real
datasets demonstrate enhanced prediction performance by capturing the dynamics
of different variables. Meanwhile, we evaluate the interpretation results both
qualitatively and quantitatively. It exhibits the prospect as an end-to-end
framework for both forecasting and knowledge extraction over multi-variable
data.Comment: Accepted to International Conference on Machine Learning (ICML), 201
DeepLOB: Deep Convolutional Neural Networks for Limit Order Books
We develop a large-scale deep learning model to predict price movements from
limit order book (LOB) data of cash equities. The architecture utilises
convolutional filters to capture the spatial structure of the limit order books
as well as LSTM modules to capture longer time dependencies. The proposed
network outperforms all existing state-of-the-art algorithms on the benchmark
LOB dataset [1]. In a more realistic setting, we test our model by using one
year market quotes from the London Stock Exchange and the model delivers a
remarkably stable out-of-sample prediction accuracy for a variety of
instruments. Importantly, our model translates well to instruments which were
not part of the training set, indicating the model's ability to extract
universal features. In order to better understand these features and to go
beyond a "black box" model, we perform a sensitivity analysis to understand the
rationale behind the model predictions and reveal the components of LOBs that
are most relevant. The ability to extract robust features which translate well
to other instruments is an important property of our model which has many other
applications.Comment: 12 pages, 9 figure
- …