145 research outputs found

    Financial distress prediction using the hybrid associative memory with translation

    Get PDF
    This paper presents an alternative technique for financial distress prediction systems. The method is based on a type of neural network, which is called hybrid associative memory with translation. While many different neural network architectures have successfully been used to predict credit risk and corporate failure, the power of associative memories for financial decision-making has not been explored in any depth as yet. The performance of the hybrid associative memory with translation is compared to four traditional neural networks, a support vector machine and a logistic regression model in terms of their prediction capabilities. The experimental results over nine real-life data sets show that the associative memory here proposed constitutes an appropriate solution for bankruptcy and credit risk prediction, performing significantly better than the rest of models under class imbalance and data overlapping conditions in terms of the true positive rate and the geometric mean of true positive and true negative rates.This work has partially been supported by the Mexican CONACYT through the Postdoctoral Fellowship Program [232167], the Spanish Ministry of Economy [TIN2013-46522-P], the Generalitat Valenciana [PROMETEOII/2014/062] and the Mexican PRODEP [DSA/103.5/15/7004]. We would like to thank the Reviewers for their valuable comments and suggestions, which have helped to improve the quality of this paper substantially

    Prediction of Banks Financial Distress

    Get PDF
    In this research we conduct a comprehensive review on the existing literature of prediction techniques that have been used to assist on prediction of the bank distress. We categorized the review results on the groups depending on the prediction techniques method, our categorization started by firstly using time factors of the founded literature, so we mark the literature founded in the period (1990-2010) as history of prediction techniques, and after this period until 2013 as recent prediction techniques and then presented the strengths and weaknesses of both. We came out by the fact that there was no specific type fit with all bank distress issue although we found that intelligent hybrid techniques considered the most candidates methods in term of accuracy and reputatio

    Predicting Bankruptcy After The Sarbanes-Oxley Act Using The Most Current Data Mining Approaches

    Get PDF
    Our study proposes several current data mining methods to predict bankruptcy after the Sarbanes-Oxley Act (2002) using 2007-2008 U.S. data.  The Sarbanes-Oxley Act (SOX) of 2002 was introduced to improve the quality of financial reporting and minimize corporate fraud in the U.S.  Because of this SOX implementation, a company’s financial statements are assumed to provide higher quality financial information for investors and other stakeholders. The results of our data mining approaches in our bankruptcy prediction study show that Bayesian Net method performs the best (85% overall prediction rate with 94% in AUC), followed by J48 (85% with 82% AUC), Decision Table (83.52%), and Decision Tree (82%) methods using financial and other data from the 10-K report and Compustat.  These results are better than previous bankruptcy prediction studies before the SOX implementation using most current data mining approaches

    Predicting Corporate Bankruptcy: Lessons from the Past

    Get PDF
    The need for corporate bankruptcy prediction models arises in 1960 after the increase in incidence of some major bankruptcies. Over the years, the episodes of financial turmoil increase in number and so does these bankruptcy prediction models. Existing reviews of bankruptcy models are either narrowly focused or outdated. Current study aims to provide an overview of the existing models for predicting bankruptcy and review the significance of these models. Furthermore, it highlights the problems and issues in the existing models which hinders the accuracy in predicting bankruptcy

    Failure prediction of European high-tech companies

    Get PDF
    The aim of this thesis is to develop a model for predicting the failure of high-tech and mediumhigh tech companies from different European countries. This study uses firm-level data from the Bureau van Dijk’s Amadeus database and includes the financial information of 32,929 firms. The data were collected from the financial statements of the companies for the period 2012–2017 and logistic regression was used as the analysis method. Findings indicate that the accuracies of individual variables across countries are not very high and there are large differences in the accuracies of individual ratios when comparing non-failed and failed firms. Aggregate accuracies for all ratios within country and across countries show that the most accurate predictions are obtained for non-failed firms using the ratios for the preceding two years combined. The practical value of this work lies in the knowledge of the relevant variables, which allows companies to focus in a timely manner on aspects that have determined failure in the past. Subsequent works should attempt to use a larger sample of European countries and include other variables in addition to financial ratios

    Assessment of trajectories of non-bankrupt and bankrupt enterprises

    Get PDF
    Purpose: The aim of this study is to show how long-term trajectories of enterprises can be used to increase the forecasting horizon of bankruptcy prediction models. Design/Methodology/Approach: The author used seven popular forecasting models (two from Europe, two from Asia, two from North America and one from Latin America). These models (five multivariate discriminant analysis models and two logit models) were used to develop 17-year trajectories separately for non-bankrupt enterprises and those at risk of financial failure. Findings: Based on a sample of 200 enterprises, the author evaluated the differences between non-bankrupt and bankrupt firms in development during 17 years of activity. The long-term usability of the models was demonstrated. To date, these models have been used only to forecast bankruptcy risk in the short term (1–3 years’ prediction horizon). This paper demonstrates that these models can also serve to evaluate long-term growth and to identify the first symptoms of future bankruptcy risk many years before it actually occurs. Practical Implications: It was proven and specified that long-term developmental differences exist between non-threatened and future insolvent companies. These studies proved that the process of going bankrupt is very long, perhaps even longer than the literature has previously demonstrated. Originality/value: This study is one of the first attempts in the literature globally to assess such long-term enterprise trajectories. Additionally by implementing a dynamic approach to the financial ratios in the risk-forecasting model let visualize the changes occurring in the company.peer-reviewe

    Data Science for Finance: Targeted Learning from (Big) Data to Economic Stability and Financial Risk Management

    Get PDF
    A thesis submitted in partial fulfillment of the requirements for the degree of Doctor in Information Management, specialization in Statistics and EconometricsThe modelling, measurement, and management of systemic financial stability remains a critical issue in most countries. Policymakers, regulators, and managers depend on complex models for financial stability and risk management. The models are compelled to be robust, realistic, and consistent with all relevant available data. This requires great data disclosure, which is deemed to have the highest quality standards. However, stressed situations, financial crises, and pandemics are the source of many new risks with new requirements such as new data sources and different models. This dissertation aims to show the data quality challenges of high-risk situations such as pandemics or economic crisis and it try to theorize the new machine learning models for predictive and longitudes time series models. In the first study (Chapter Two) we analyzed and compared the quality of official datasets available for COVID-19 as a best practice for a recent high-risk situation with dramatic effects on financial stability. We used comparative statistical analysis to evaluate the accuracy of data collection by a national (Chinese Center for Disease Control and Prevention) and two international (World Health Organization; European Centre for Disease Prevention and Control) organizations based on the value of systematic measurement errors. We combined excel files, text mining techniques, and manual data entries to extract the COVID-19 data from official reports and to generate an accurate profile for comparisons. The findings show noticeable and increasing measurement errors in the three datasets as the pandemic outbreak expanded and more countries contributed data for the official repositories, raising data comparability concerns and pointing to the need for better coordination and harmonized statistical methods. The study offers a COVID-19 combined dataset and dashboard with minimum systematic measurement errors and valuable insights into the potential problems in using databanks without carefully examining the metadata and additional documentation that describe the overall context of data. In the second study (Chapter Three) we discussed credit risk as the most significant source of risk in banking as one of the most important sectors of financial institutions. We proposed a new machine learning approach for online credit scoring which is enough conservative and robust for unstable and high-risk situations. This Chapter is aimed at the case of credit scoring in risk management and presents a novel method to be used for the default prediction of high-risk branches or customers. This study uses the Kruskal-Wallis non-parametric statistic to form a conservative credit-scoring model and to study its impact on modeling performance on the benefit of the credit provider. The findings show that the new credit scoring methodology represents a reasonable coefficient of determination and a very low false-negative rate. It is computationally less expensive with high accuracy with around 18% improvement in Recall/Sensitivity. Because of the recent perspective of continued credit/behavior scoring, our study suggests using this credit score for non-traditional data sources for online loan providers to allow them to study and reveal changes in client behavior over time and choose the reliable unbanked customers, based on their application data. This is the first study that develops an online non-parametric credit scoring system, which can reselect effective features automatically for continued credit evaluation and weigh them out by their level of contribution with a good diagnostic ability. In the third study (Chapter Four) we focus on the financial stability challenges faced by insurance companies and pension schemes when managing systematic (undiversifiable) mortality and longevity risk. For this purpose, we first developed a new ensemble learning strategy for panel time-series forecasting and studied its applications to tracking respiratory disease excess mortality during the COVID-19 pandemic. The layered learning approach is a solution related to ensemble learning to address a given predictive task by different predictive models when direct mapping from inputs to outputs is not accurate. We adopt a layered learning approach to an ensemble learning strategy to solve the predictive tasks with improved predictive performance and take advantage of multiple learning processes into an ensemble model. In this proposed strategy, the appropriate holdout for each model is specified individually. Additionally, the models in the ensemble are selected by a proposed selection approach to be combined dynamically based on their predictive performance. It provides a high-performance ensemble model to automatically cope with the different kinds of time series for each panel member. For the experimental section, we studied more than twelve thousand observations in a portfolio of 61-time series (countries) of reported respiratory disease deaths with monthly sampling frequency to show the amount of improvement in predictive performance. We then compare each country’s forecasts of respiratory disease deaths generated by our model with the corresponding COVID-19 deaths in 2020. The results of this large set of experiments show that the accuracy of the ensemble model is improved noticeably by using different holdouts for different contributed time series methods based on the proposed model selection method. These improved time series models provide us proper forecasting of respiratory disease deaths for each country, exhibiting high correlation (0.94) with Covid-19 deaths in 2020. In the fourth study (Chapter Five) we used the new ensemble learning approach for time series modeling, discussed in the previous Chapter, accompany by K-means clustering for forecasting life tables in COVID-19 times. Stochastic mortality modeling plays a critical role in public pension design, population and public health projections, and in the design, pricing, and risk management of life insurance contracts and longevity-linked securities. There is no general method to forecast the mortality rate applicable to all situations especially for unusual years such as the COVID-19 pandemic. In this Chapter, we investigate the feasibility of using an ensemble of traditional and machine learning time series methods to empower forecasts of age-specific mortality rates for groups of countries that share common longevity trends. We use Generalized Age-Period-Cohort stochastic mortality models to capture age and period effects, apply K-means clustering to time series to group countries following common longevity trends, and use ensemble learning to forecast life expectancy and annuity prices by age and sex. To calibrate models, we use data for 14 European countries from 1960 to 2018. The results show that the ensemble method presents the best robust results overall with minimum RMSE in the presence of structural changes in the shape of time series at the time of COVID-19. In this dissertation’s conclusions (Chapter Six), we provide more detailed insights about the overall contributions of this dissertation on the financial stability and risk management by data science, opportunities, limitations, and avenues for future research about the application of data science in finance and economy

    Calculando el riesgo de insolvencia, de los métodos tradicionales a las redes neuronales artificiales. Una revisión de literatura

    Get PDF
    In the financial management of any organization, the calculation of the risk of insolvency has become an important parameter, seeking to anticipate the eventuality of having an economic problem and generating insolvency. The objective of this work is to compare the classical methodologies and the artificial neural networks applied to calculate the risk of insolvency, looking for the main characteristics within the applications carried out by different authors over time. In this way, the main variables that can show that the application of the neural network methodology facilitates the calculation of the risk of insolvency are observed. Through the bibliographic review, between the years 1992-2021, with the use of the analytical-synthetic method, it can be seen that the exposed model is considered efficient according to its results, with adjustments that, in most of the exposed cases, exceed 80% efficiency. The results found allowed us to conclude that the basic structure of a neural network is given by three layers: an input layer, an output layer and a hidden layer. However, the number of nodes varies in each of the applications carried out by the different authors, since they represent the variables, in this case the most relevant financial indicators according to the proposed application. Finally, it was possible to show which are the most used financial indicators in the different neural network applications.En la administración financiera de toda organización el cálculo del riesgo de insolvencia se ha convertido en un parámetro importante, buscando anticiparse a la eventualidad de llegar a tener un problema económico y generar insolvencia. El objetivo de este trabajo es determinar si en el cálculo del riesgo de insolvencia, el uso de redes neuronales artificiales genera mejores resultados que las metodologías tradicionales, buscando las principales características dentro de las aplicaciones realizadas por distintos autores a través del tiempo. De esta manera se observan las principales variables que pueden evidenciar que la aplicación de la metodología de redes neuronales facilita el cálculo del riesgo de insolvencia.  A través de la revisión bibliográfica, en el período 1992-2021, con el uso del método analítico-sintético se puede evidenciar que el modelo expuesto es considerado como eficiente según sus resultados, con ajustes que, en la mayoría de los casos expuestos, superan el 80% de eficacia. Los resultados encontrados permitieron concluir que la estructura básica de una red neuronal viene dada por tres capas: una de entrada, una de salida y una oculta. Sin embargo, el número de nodos es el que varía en cada una de las aplicaciones realizadas por los distintos autores, dado que los mismos representan a las variables, en este caso indicadores financieros más relevantes según la aplicación planteada. Finalmente se logró evidenciar cuáles son los indicadores financieros más usados en las distintas aplicaciones de redes neuronales. Todo indica que las redes neuronales generan resultados más efectivos que los métodos tradicionales.&nbsp
    corecore