128,820 research outputs found
Time series prediction and forecasting using Deep learning Architectures
Nature brings time series data everyday and everywhere, for example, weather data, physiological signals and biomedical signals, financial and business recordings. Predicting the future observations of a collected sequence of historical observations is called time series forecasting. Forecasts are essential, considering the fact that they guide decisions in many areas of scientific, industrial and economic activity such as in meteorology, telecommunication, finance, sales and stock exchange rates. A massive amount of research has already been carried out by researchers over many years for the development of models to improve the time series forecasting accuracy. The major aim of time series modelling is to scrupulously examine the past observation of time series and to develop an appropriate model which elucidate the inherent behaviour and pattern existing in time series. The behaviour and pattern related to various time series may possess different conventions and infact requires specific countermeasures for modelling. Consequently, retaining the neural networks to predict a set of time series of mysterious domain remains particularly challenging. Time series forecasting remains an arduous problem despite the fact that there is substantial improvement in machine learning approaches. This usually happens due to some factors like, different time series may have different flattering behaviour. In real world time series data, the discriminative patterns residing in the time series are often distorted by random noise and affected by high-frequency perturbations. The major aim of this thesis is to contribute to the study and expansion of time series prediction and multistep ahead forecasting method based on deep learning algorithms. Time series forecasting using deep learning models is still in infancy as compared
to other research areas for time series forecasting.Variety of time series data has been considered in this research. We explored several deep learning architectures on
the sequential data, such as Deep Belief Networks (DBNs), Stacked AutoEncoders (SAEs), Recurrent Neural Networks (RNNs) and Convolutional Neural Networks
(CNNs). Moreover, we also proposed two different new methods based on muli-step ahead forecasting for time series data. The comparison with state of the art methods is also exhibited. The research work conducted in this thesis makes theoretical, methodological and empirical contributions to time series prediction and multi-step ahead forecasting by using Deep Learning Architectures
λ€μ°¨μ κ±°λ λ°μ΄ν°μ λ₯λ¬λ λͺ¨νμ μ΄μ©ν κΈμ΅ μμ₯ 볡μ‘κ³ λ€νΈμν¬ λΆμ
νμλ
Όλ¬Έ(λ°μ¬) -- μμΈλνκ΅λνμ : 곡과λν μ°μ
곡νκ³Ό, 2021.8. μ₯μ°μ§.μ£Όμμ λ
립μ μμκ° μλ, λ€λ₯Έ μμ°λ€μ ν¬ν¨ν λ€μν μμλ€κ³Ό μνΈ μ°κ΄μ±μ΄ μλ€. λ°λΌμ μ£Όμμ μνΈ μ°κ΄μ±μ λΆμνλ κ²μ μ£Όμ μμ₯μ ꡬ쑰λ₯Ό λ¬μ¬νκ³ μ΄ν΄νκΈ° μν νμμ μμμΌ λΏλ§ μλλΌ, ν¬μμ μνκ΄λ¦¬μ κ°μ μ€λ¬΄μ κ΄μ μμλ μ€μν μμμ΄λ€.
μ£Όμ μμ₯μ κ°λ³ μμΈμ μνΈ μμ©μ ν΅ν΄ νμ±λ 볡μ‘κ³ λ€νΈμν¬λΌλ κ΄μ νμ, λ§μ μ°κ΅¬λ€μ΄ μ£Όμ κ°μ μνΈ μ°κ΄μ±μ λΆμνκΈ° μν΄ κ²½μ 물리νμ λ€νΈμν¬ λΆμμ μ¬μ©νκ³ μλ€.
μ£Όμ κ°μ μνΈ κ΄κ³λ₯Ό μ μ ν λ, λ§μ μ°κ΅¬μμ μ£Όκ° μκ³μ΄μλ§ μ΄μ μ λλ€.
κ·Έλ¬λ, μ£Όμμ λνλ΄λ μ 보λ μ£Όκ°λΏλ§ μλλΌ κ±°λλ, λ³λμ±, κ·Έλ¦¬κ³ νκ³ μ 보μ κ°μ λ€μν λ°μ΄ν°κ° μλ€.
λ°λΌμ μ£Όκ°λ§μ μ΄μ©νμ¬ μ£Όμ κ°μ μνΈ μ°κ΄μ±μ λΆμνλ κ²μ, κΈμ΅ μμ₯μ λν λ¨νΈμ μΈ μκ°λ§μ μλ―Έν μ μλ€.
κ²λ€κ°, λ₯λ¬λκ³Ό κ°μ κΈ°μ λ ₯μ λ°μ μ λ°λΌ, νλ¬Έκ³Ό λ°μ΄ν°μ μ’
λ₯μ κ²½κ³λ₯Ό ν무λ νμ κ° μ΅ν©μ μ€μμ±μ΄ λλλκ³ μλ€.
μ΄λ¬ν λ§₯λ½μμ λ³Έ λ
Όλ¬Έμ λͺ©μ μ κ²½μ 물리νκ³Ό λ₯λ¬λμ νμ κ° μ΅ν©μ ν΅ν΄ κΈμ΅ μμ₯ λΆμμ λν μλ‘μ΄ μκ°μ μ μνλλ° μλ€.
μ°λ¦¬λ S\&P500 μ£Όμ μμ₯μ λ€λ³λ κ΄μ μμ λΆμνκΈ° μν΄ μ΄μ°¨μ νμ€ν κ·Έλ¨(Bi-dimensional histogram)κ³Ό μ€ν μΈμ½λ(Autoencoder)μ κΈ°λ°ν λ₯λ¬λκ³Ό κ²½μ 물리νμ΄ μ΅ν©λ νλ μμν¬(Framework)λ₯Ό μ μνλ€.
λ¨Όμ , μ°λ¦¬λ μ£Όμμ λ€λ³λμΌλ‘ νννκΈ° μν΄ μ£Όκ°μ κ±°λλμ ν΅ν΄ ꡬμ±λ μ΄μ°¨μ νμ€ν κ·Έλ¨μ μ μνλ€.
κ° μ£Όμ λ³ μ΄μ°¨μ νμ€ν κ·Έλ¨μ μΌλ³ μ£Όκ° μκ³μ΄κ³Ό κ±°λλ μκ³μ΄λ‘ ꡬμ±λλ©°, μ£Όμμ λΆν¬ κ΄μ μμ μ§κ΄μ μΌλ‘ νννλλ° λͺ©μ μ΄ μλ€.
κ·Έ λ€μ, μ΄μ°¨μ νμ€ν κ·Έλ¨μ μ°¨μμ μΆμνκ³ μλ―Έμλ λ³μλ₯Ό μΆμΆνκΈ° μν΄ μ€ν μΈμ½λκ° μ μ©λλ€.
μΆμΆλ λ³μλ₯Ό ν΅ν΄ μ΄μ°¨μ νμ€ν κ·Έλ¨ κ°μ 거리 νλ ¬(Distance matrix)μ΄ μ μλκ³ , 거리νλ ¬μ κ²½μ 물리ν λ°©λ²λ‘ μΈ PMFG(Planar Maximally Filtered Graph) μκ³ λ¦¬μ¦μ μ μ©νμ¬ κΈμ΅ μμ₯μ λνλ΄λ νμ€ν κ·Έλ¨ λ€νΈμν¬κ° ꡬμΆλλ€.
ꡬμΆλ λ€νΈμν¬λ μ΄μ°¨μ νμ€ν κ·Έλ¨μ μ μ¬ κ³΅κ°(Latent space)λ₯Ό μλ―Ένλ©°, κΈμ΅ μμ₯μ ꡬ쑰μ νΉμ±μ λΆμνκΈ° μν λ€νΈμν¬ λΆμμ΄ μνλλ€.
μ°λ¦¬λ νμ€ν κ·Έλ¨μ κΈ°λ°ν λ€νΈμν¬μ ꡬ쑰μ νΉμ±μ΄ νμ€ν κ·Έλ¨μ μ°ν¬λμ μ°κ΄μ΄ μμμ λ°κ²¬νμλ€. μ΄λ₯Ό ν΅ν΄ μ°λ¦¬λ μ€ν μΈμ½λκ° μ΄μ°¨μ νμ€ν κ·Έλ¨μ λ΄μ¬λ λ³μλ₯Ό μΆμΆνλλ° ν¨κ³Όμ μμ νμΈνμλ€.
λν μ°λ¦¬λ νμ€ν κ·Έλ¨ λ€νΈμν¬λ κΈ°μ‘΄ λ°©λ²λ‘ μΈ μ£Όκ° μκ³μ΄ κΈ°λ°μ λ€νΈμν¬μ μμνμ νΉμ§μ΄ μλ‘ μμ΄ν¨μ λ°κ²¬νλ€.
μΆκ°μ μΌλ‘, νμ€ν κ·Έλ¨ λ€νΈμν¬λ κΈ°μ‘΄ λ°©λ²λ‘ λ³΄λ€ λ³λμ±κ³Ό μμ₯ λ² ν(Beta)μ κ°μ κΈμ΅ μμ₯μ λν μΈμμ μΈ νΉμ±μ΄ μ μ¬ν μ£Όμλ€μ ꡬ쑰μ μΌλ‘ κ΅°μ§ννλλ° λ ν¨κ³Όκ° μλ€.
λ λμκ°, μ°λ¦¬λ νμ€ν κ·Έλ¨ λ€νΈμν¬μ κΈ°λ°ν κ·Έλν λ΄λ΄ λ€νΈμν¬(Graph Neural Network) μ£Όκ° μΆμΈ μμΈ‘ λͺ¨νμ μ μνλ€.
μ£Όκ° μΆμΈ μμΈ‘μ μ€λ¬΄μ μ€μμ±μ λ°νμΌλ‘, κ³λμ λͺ¨νλΆν° κΈ°κ³νμ΅ λͺ¨νκ³Ό λ₯λ¬λ λͺ¨νκΉμ§ κ΄λ²μνκ² μ°κ΅¬λκ³ μλ μ£Όμ μ΄λ€.
μ΅κ·Ό κ·Έλν λ΄λ΄ λ€νΈμν¬ λͺ¨νμ μμ κ°μ μνΈ μ°κ΄μ±μ λ°μ ν μ μλ νΉμ§μ λ°νμΌλ‘ μκ³μ΄ μμΈ‘ λΆμΌμμ μ£Όλͺ© λ°κ³ μλ€.
μ΅κ·Ό, μ¬λ¬ μ°κ΅¬λ€μ΄ κ·Έλν λ΄λ΄ λ€νΈμν¬μ μ£Όμ κ°μ μνΈ μ°κ΄μ±μ λ°μνμ¬ μ£Όκ° μΆμΈ μμΈ‘μ λν μ°κ΅¬λ₯Ό μ§ννλ€.
νμ§λ§, μ΄λ¬ν μ ν μ°κ΅¬λ€μ μ£Όμ κ°μ μνΈ μ°κ΄μ±μ μ 립νλλ° μ°μ
ꡬ쑰, λΉμ¦λμ€ μμκ³Ό κ°μ μ μ μ΄κ³ μ¬μ μ μκ° μꡬλλ κ΄κ³λ§μ μ΄μ©νλ€.
νμ§λ§ λ§μ μ°κ΅¬λ€μ ν΅ν΄ κΈμ΅ μμ₯μ ꡬ쑰 λ° μ£Όμ κ°μ μνΈ μ°κ΄μ±μ μλ³ μμ±(Time-varying property)μ΄ μ
μ¦λμλ€.
λ°λΌμ μ°λ¦¬λ κΈμ΅ μμ₯μ μλ³ μμ±μ λͺ¨νμ λ°μνκ³ μ, νμ€ν κ·Έλ¨ λ€νΈμν¬λ₯Ό μ£Όμ κ° μνΈ κ΄κ³λ‘ νμ©νλ κ·Έλν λ΄λ΄ λ€νΈμν¬ λͺ¨νμ μ μνλ€.
μ£Όκ° μΆμΈ μμΈ‘ μ€νμ 10μΌκ³Ό 20μΌ λ€μ μ£Όκ°μ λ°©ν₯ λΆλ₯(Classification) λ¬Έμ λ‘ μ΄λ£¨μ΄μ§λ€.
μ€ν κ²°κ³Ό, μ°λ¦¬λ λ³Έ νμ λ
Όλ¬Έμμ μ μνλ λͺ¨νμ΄ λ²€μΉλ§ν¬(Benchmark) λλΉ μ°μν μμΈ‘ μ νλμ, F1-Scoreλ₯Ό κ°λ κ²μ νμΈνλ€.
λ§μ§λ§μΌλ‘ μ°λ¦¬λ μ£Όκ° μΆμΈ μμΈ‘ λͺ¨νμ μ€λ¬΄μ νμ©λλ₯Ό κ²μ¦νκΈ° μν΄ μμΈ‘ λͺ¨νμ κΈ°λ°ν ν¬μ μ λ΅μ μ μνλ€.
μ μνλ ν¬μ μ λ΅μ μ 체 μ£Όμμ λμμΌλ‘ μΆμΈ μμΈ‘μ μννμ λ, μμΉ νλ₯ μ΄ κ°μ₯ λμ κ°μ μ£Όμμ ν¬μ ν¬νΈν΄λ¦¬μ€λ‘ ꡬμ±νλ κ²μ΄λ€.
μ¦, μ°λ¦¬λ νμ€ν κ·Έλ¨ λ€νΈμν¬μ κΈ°λ°ν κ·Έλν λ΄λ΄ λ€νΈμν¬ λͺ¨νμΌλ‘ μΆμΈ μμΈ‘μ μννκ³ , μμΈ‘ κ²°κ³Όλ₯Ό νμ©νμ¬ ν¬μ ν¬νΈν΄λ¦¬μ€λ₯Ό ꡬμ±νλ€.
ν¬μ μ€ν κ²°κ³Ό, μ μνλ ν¬μ μ λ΅μ μ ν΅μ ν¬μ μ λ΅ λλΉ μ€ν μ§μ(Sharpe ratio), μλ₯΄ν°λ
Έ μ§μ(Sortino ratio)μ κ°μ μμ΅μ± μ§νκ° λ°μ΄λ μ λ΅μΌ λΏλ§ μλλΌ, νλ½ μΆμΈμμ κ°μ₯ λΉ λ₯΄κ² λ²μ΄λ μ μλ ν¬μ μ λ΅μ΄λ€.
μ 리νλ©΄, λ³Έ νμ λ
Όλ¬Έμ κ²½μ 물리νκ³Ό λ₯λ¬λμ μ΅ν©μ κΈ°λ°ν κΈμ΅ μμ₯ ꡬ쑰ν λ°©λ²λ‘ λ° μμΈ‘ λͺ¨λΈμ μ μνλλ° μμκ° μμΌλ©°, κ΅°μ§ν, μμΈ‘, κ·Έλ¦¬κ³ ν¬μλ₯Ό ν΅ν΄ κ·Έ μ μ©μ±μ κ²μ¦νμλ€.Stocks are correlated with a variety of factors, such as other financial assets and economic factors.
Therefore, analyzing the relationship of stocks is an essential element for understanding the stock market structure and an essential element from a practical point of view, such as investment and risk management.
Stock markets that fluctuate chaotically are modeled as complex networks formed through the interaction of individual factors.
Many studies have used network analysis in econophysics to investigate the relationship between stocks.
In defining interactions between stocks, many studies highlighted stock price time series.
However, information on stocks is not only contained in the price, but also in various data such as trading volumes, volatility, and accounting variables.
Therefore, the relationship between stocks using only the stock price can show a fragmentary view of the financial market.
In addition, with the development of technologies such as deep learning, the boundaries between disciplines and data types are collapsing, and interdisciplinary convergence is becoming more significant.
In this context, the purpose of this dissertation is to present a new perspective on financial market analysis through the interdisciplinary convergence of econophysics and deep learning.
We propose a framework that combines deep learning and econophysics consisting of a bi-dimensional histogram and an autoencoder to analyze the S\&P500 stock market from a multivariate perspective.
We utilize a bi-dimensional histogram to intuitively represent stock trading volume information as well as stock price information.
Then, an autoencoder is applied to reduce the dimension of the bi-dimensional histogram and extract the latent vector.
A distance matrix between bi-dimensional histograms is defined through the latent vectors, and a histogram network representing the financial market is constructed by applying the Planar Maximally Filtered Graph(PMFG) algorithm, an econophysics methodology, to the distance matrix.
The constructed network implies the latent space of the bi-dimensional histogram, and network analysis is performed to analyze the structural properties of the financial market.
We reveal that the structural properties of the histogram network are related to the dispersion of histogram, which means that the autoencoder is effective in extracting the latent vector of the bi-dimensional histogram.
Additionally, the histogram network effectively clusters stocks with similar exogenous characteristics to financial markets, such as volatility and market beta.
Furthermore, we propose a stock price prediction model based on a Graph Neural Network(GNN) and histogram network.
Based on its practical importance, stock price trend prediction is a research topic studied extensively, from econometric models to machine learning and deep learning.
Recently, GNN models are in the spotlight in time series prediction based on their properties that can reflect the relationship between elements.
However, previous studies use only static and predefined relationships such as industrial structures to establish the relationship between stocks.
However, many studies have demonstrated the time-varying property of the financial market.
Therefore, we propose a graph neural network model that uses the histogram network as a relationship between stocks to reflect the time-varying properties of the financial market.
The trend prediction experiment is performed with a classification problem of the stock price after 10 and 20 days.
As a result, we confirmed that the proposed model has superior prediction accuracy and F1-Score compared to the benchmark.
Finally, we propose an investment strategy based on the predictive model to verify the practical utility of the proposed model.
The proposed investment strategy is to construct an investment portfolio of stocks that are most likely to rise in price in the future.
In other words, we perform trend prediction with a graph neural network model based on a histogram network and construct an investment portfolio using the prediction results.
Our results show that the proposed investment strategy is outstanding in profitability, including the Sharpe ratio and Sortino ratio and the cumulative return.
In addition, our proposed investment strategy can recover from a downtrend faster than traditional investment strategies based on Modern Portfolio Theory.
In summary, the novelty of the dissertation is that we construct a stock market PMFG network and build a GNN model based on the interdisciplinary convergence of econophysics and deep learning. The usefulness of the dissertation is verified through clustering, prediction, and investment experiments.Chapter 1 Introduction 1
1.1 Research Motivation and Purpose 1
1.2 Organization of the Research 8
Chapter 2 Literature Review 11
2.1 Stock market network 11
2.2 Autoencoder in Stock market 12
2.3 Predictive model in Stock market 13
Chapter 3 Bi-dimensional Histogram Network 17
3.1 Architecture of Bi-dimensional Histogram Network 17
3.1.1 Bi-dimensional Histogram 17
3.1.2 Auto-encoder 21
3.1.3 Stock market PMFG network 25
3.2 Data description 28
3.3 Experimental Results 29
3.3.1 Properties of bi-dimensional histogram 29
3.3.2 Dimensional reduction result of bi-dimensional histogram 34
3.3.3 Properties of bi-dimensional histogram network 36
3.3.4 Comparison of histogram network with price network 40
3.4 Summary and Discussion 47
Chapter 4 A hybrid graph neural network for time series trend prediction 49
4.1 Proposed Model 49
4.2 Benchmark Algorithms 58
4.3 Data Description 63
4.4 Experimental Procedure 64
4.5 Experimental Result 68
4.5.1 Properties of High-Low volatility PMFG 68
4.5.2 Prediction Performance 74
4.5.3 Robustness Test for Prediction Performance 79
4.6 Summary and Discussion 82
Chapter 5 Investment Strategy with Trend Prediction Model 85
5.1 Proposed Strategy 85
5.2 Benchmark Algorithms 87
5.3 Experiment Procedure 91
5.4 Experimental Result 94
5.5 Summary and Discussion 110
Chapter 6 Conclusion 113
6.1 Conclusions 113
6.2 Future Works 117
Appendix 119
Bibliography 129
κ΅λ¬Έμ΄λ‘ 145λ°
DeepLOB: Deep Convolutional Neural Networks for Limit Order Books
We develop a large-scale deep learning model to predict price movements from
limit order book (LOB) data of cash equities. The architecture utilises
convolutional filters to capture the spatial structure of the limit order books
as well as LSTM modules to capture longer time dependencies. The proposed
network outperforms all existing state-of-the-art algorithms on the benchmark
LOB dataset [1]. In a more realistic setting, we test our model by using one
year market quotes from the London Stock Exchange and the model delivers a
remarkably stable out-of-sample prediction accuracy for a variety of
instruments. Importantly, our model translates well to instruments which were
not part of the training set, indicating the model's ability to extract
universal features. In order to better understand these features and to go
beyond a "black box" model, we perform a sensitivity analysis to understand the
rationale behind the model predictions and reveal the components of LOBs that
are most relevant. The ability to extract robust features which translate well
to other instruments is an important property of our model which has many other
applications.Comment: 12 pages, 9 figure
Enhanced news sentiment analysis using deep learning methods
We explore the predictive power of historical news sentiments based on financial market performance to forecast financial news sentiments. We define news sentiments based on stock price returns averaged over one minute right after a news article has been released. If the stock price exhibits positive (negative) return, we classify the news article released just prior to the observed stock return as positive (negative). We use Wikipedia and Gigaword five corpus articles from 2014 and we apply the global vectors for word representation method to this corpus to create word vectors to use as inputs into the deep learning TensorFlow network. We analyze high-frequency (intraday) Thompson Reuters News Archive as well as the high-frequency price tick history of the Dow Jones Industrial Average (DJIA 30) Index individual stocks for the period between 1/1/2003 and 12/30/2013. We apply a combination of deep learning methodologies of recurrent neural network with long short-term memory units to train the Thompson Reuters News Archive Data from 2003 to 2012, and we test the forecasting power of our method on 2013 News Archive data. We find that the forecasting accuracy of our methodology improves when we switch from random selection of positive and negative news to selecting the news with highest positive scores as positive news and news with highest negative scores as negative news to create our training data set.Published versio
- β¦