8,959 research outputs found
An academic review: applications of data mining techniques in finance industry
With the development of Internet techniques, data volumes are doubling every two years, faster than predicted by Moore’s Law. Big Data Analytics becomes particularly important for enterprise business. Modern computational technologies will provide effective tools to help understand hugely accumulated data and leverage this information to get insights into the finance industry. In order to get actionable insights into the business, data has become most valuable asset of financial organisations, as there are no physical products in finance industry to manufacture. This is where data mining techniques come to their rescue by allowing access to the right information at the right time. These techniques are used by the finance industry in various areas such as fraud detection, intelligent forecasting, credit rating, loan management, customer profiling, money laundering, marketing and prediction of price movements to name a few. This work aims to survey the research on data mining techniques applied to the finance industry from 2010 to 2015.The review finds that Stock prediction and Credit rating have received most attention of researchers, compared to Loan prediction, Money Laundering and Time Series prediction. Due to the dynamics, uncertainty and variety of data, nonlinear mapping techniques have been deeply studied than linear techniques. Also it has been proved that hybrid methods are more accurate in prediction, closely followed by Neural Network technique. This survey could provide a clue of applications of data mining techniques for finance industry, and a summary of methodologies for researchers in this area. Especially, it could provide a good vision of Data Mining Techniques in computational finance for beginners who want to work in the field of computational finance
Credit scoring: comparison of non‐parametric techniques against logistic regression
Dissertation presented as the partial requirement for obtaining a Master's degree in Information Management, specialization in Knowledge Management and Business IntelligenceOver the past decades, financial institutions have been giving increased importance to credit risk
management as a critical tool to control their profitability. More than ever, it became crucial for
these institutions to be able to well discriminate between good and bad clients for only
accepting the credit applications that are not likely to default. To calculate the probability of
default of a particular client, most financial institutions have credit scoring models based on
parametric techniques. Logistic regression is the current industry standard technique in credit
scoring models, and it is one of the techniques under study in this dissertation. Although it is
regarded as a robust and intuitive technique, it is still not free from several critics towards the
model assumptions it takes that can compromise its predictions. This dissertation intends to
evaluate the gains in performance resulting from using more modern non-parametric
techniques instead of logistic regression, performing a model comparison over four different
real-life credit datasets. Specifically, the techniques compared against logistic regression in this
study consist of two single classifiers (decision tree and SVM with RBF kernel) and two ensemble
methods (random forest and stacking with cross-validation). The literature review demonstrates
that heterogeneous ensemble approaches have a weaker presence in credit scoring studies and,
because of that, stacking with cross-validation was considered in this study. The results
demonstrate that logistic regression outperforms the decision tree classifier, has similar
performance in relation to SVM and slightly underperforms both ensemble approaches in similar
extents
Revealing Robust Oil and Gas Company Macro-Strategies using Deep Multi-Agent Reinforcement Learning
The energy transition potentially poses an existential risk for major
international oil companies (IOCs) if they fail to adapt to low-carbon business
models. Projections of energy futures, however, are met with diverging
assumptions on its scale and pace, causing disagreement among IOC
decision-makers and their stakeholders over what the business model of an
incumbent fossil fuel company should be. In this work, we used deep multi-agent
reinforcement learning to solve an energy systems wargame wherein players
simulate IOC decision-making, including hydrocarbon and low-carbon investments
decisions, dividend policies, and capital structure measures, through an
uncertain energy transition to explore critical and non-linear governance
questions, from leveraged transitions to reserve replacements. Adversarial play
facilitated by state-of-the-art algorithms revealed decision-making strategies
robust to energy transition uncertainty and against multiple IOCs. In all
games, robust strategies emerged in the form of low-carbon business models as a
result of early transition-oriented movement. IOCs adopting such strategies
outperformed business-as-usual and delayed transition strategies regardless of
hydrocarbon demand projections. In addition to maximizing value, these
strategies benefit greater society by contributing substantial amounts of
capital necessary to accelerate the global low-carbon energy transition. Our
findings point towards the need for lenders and investors to effectively
mobilize transition-oriented finance and engage with IOCs to ensure responsible
reallocation of capital towards low-carbon business models that would enable
the emergence of fossil fuel incumbents as future low-carbon leaders
Critical Market Crashes
This review is a partial synthesis of the book ``Why stock market crash''
(Princeton University Press, January 2003), which presents a general theory of
financial crashes and of stock market instabilities that his co-workers and the
author have developed over the past seven years. The study of the frequency
distribution of drawdowns, or runs of successive losses shows that large
financial crashes are ``outliers'': they form a class of their own as can be
seen from their statistical signatures. If large financial crashes are
``outliers'', they are special and thus require a special explanation, a
specific model, a theory of their own. In addition, their special properties
may perhaps be used for their prediction. The main mechanisms leading to
positive feedbacks, i.e., self-reinforcement, such as imitative behavior and
herding between investors are reviewed with many references provided to the
relevant literature outside the confine of Physics. Positive feedbacks provide
the fuel for the development of speculative bubbles, preparing the instability
for a major crash. We demonstrate several detailed mathematical models of
speculative bubbles and crashes. The most important message is the discovery of
robust and universal signatures of the approach to crashes. These precursory
patterns have been documented for essentially all crashes on developed as well
as emergent stock markets, on currency markets, on company stocks, and so on.
The concept of an ``anti-bubble'' is also summarized, with two forward
predictions on the Japanese stock market starting in 1999 and on the USA stock
market still running. We conclude by presenting our view of the organization of
financial markets.Comment: Latex 89 pages and 38 figures, in press in Physics Report
Macroeconomic Policy in DGSE and Agent-Based Models Redux:New Developments and Challenges Ahead
The Great Recession seems to be a natural experiment for economic analysis, in that it has shown the inadequacy of the predominant theoretical framework | the New Neoclassical Synthesis (NNS) | grounded on the DSGE model. In this paper, we present a critical discussion of the theoretical, empirical and political-economy pitfalls of the DSGE-based approach to policy analysis. We suggest that a more fruitful research avenue should escape the strong theoretical requirements of NNS models (e.g., equilibrium, rationality, representative agent, etc.) and consider the economy as a complex evolving system, i.e. as an ecology populated by heterogeneous agents, whose far-from-equilibrium interactions continuously change the structure of the system. This is indeed the methodological core of agent-based computational economics (ACE), which is presented in this paper. We also discuss how ACE has been applied to policy analysis issues, and we provide a survey of macroeconomic policy applications ( fiscal and monetary policy, bank regulation, labor market structural reforms and climate change interventions). Finally, we conclude by discussing the methodological status of ACE, as well as the problems it raises
A Bayesian Perspective of Statistical Machine Learning for Big Data
Statistical Machine Learning (SML) refers to a body of algorithms and methods
by which computers are allowed to discover important features of input data
sets which are often very large in size. The very task of feature discovery
from data is essentially the meaning of the keyword `learning' in SML.
Theoretical justifications for the effectiveness of the SML algorithms are
underpinned by sound principles from different disciplines, such as Computer
Science and Statistics. The theoretical underpinnings particularly justified by
statistical inference methods are together termed as statistical learning
theory.
This paper provides a review of SML from a Bayesian decision theoretic point
of view -- where we argue that many SML techniques are closely connected to
making inference by using the so called Bayesian paradigm. We discuss many
important SML techniques such as supervised and unsupervised learning, deep
learning, online learning and Gaussian processes especially in the context of
very large data sets where these are often employed. We present a dictionary
which maps the key concepts of SML from Computer Science and Statistics. We
illustrate the SML techniques with three moderately large data sets where we
also discuss many practical implementation issues. Thus the review is
especially targeted at statisticians and computer scientists who are aspiring
to understand and apply SML for moderately large to big data sets.Comment: 26 pages, 3 figures, Review pape
Trustworthy Reinforcement Learning Against Intrinsic Vulnerabilities: Robustness, Safety, and Generalizability
A trustworthy reinforcement learning algorithm should be competent in solving
challenging real-world problems, including {robustly} handling uncertainties,
satisfying {safety} constraints to avoid catastrophic failures, and
{generalizing} to unseen scenarios during deployments. This study aims to
overview these main perspectives of trustworthy reinforcement learning
considering its intrinsic vulnerabilities on robustness, safety, and
generalizability. In particular, we give rigorous formulations, categorize
corresponding methodologies, and discuss benchmarks for each perspective.
Moreover, we provide an outlook section to spur promising future directions
with a brief discussion on extrinsic vulnerabilities considering human
feedback. We hope this survey could bring together separate threads of studies
together in a unified framework and promote the trustworthiness of
reinforcement learning.Comment: 36 pages, 5 figure
- …