70 research outputs found

    An empirical study on credit evaluation of SMEs based on detailed loan data

    Get PDF
    Small and micro-sized Enterprises (SMEs) are an important part of Chinese economic system.The establishment of credit evaluating model of SMEs can effectively help financial intermediaries to reveal credit risk of enterprises and reduce the cost of enterprises information acquisition. Besides it can also serve as a guide to investors which also helps companies with good credit. This thesis conducts an empirical study based on loan data from a Chinese bank of loans granted to SMEs. The study aims to develop a data-driven model that can accurately predict if a given loan has an acceptable risk from the bank’s perspective, or not. Furthermore, we test different methods to deal with the problem of unbalanced class and uncredible sample. Lastly, the importance of variables is analyzed. Remaining Unpaid Principal, Floating Interest Rate, Time Until Maturity Date, Real Interest Rate, Amount of Loan all have significant effects on the final result of the prediction.The main contribution of this study is to build a credit evaluation model of small and micro enterprises, which not only helps commercial banks accurately identify the credit risk of small and micro enterprises, but also helps to overcome creditdifficulties of small and micro enterprises.As pequenas e microempresas constituem uma parte importante do sistema econĂłmico chinĂȘs. A definição de um modelo de avaliação de crĂ©dito para estas empresas pode ajudar os intermediĂĄrios financeiros a revelarem o risco de crĂ©dito das empresas e a reduzirem o custo de aquisição de informação das empresas. AlĂ©m disso, pode igualmente servir como guia para os investidores, auxiliando tambĂ©m empresas com bom crĂ©dito. Na presente tese apresenta-se um estudo empĂ­rico baseado em dados de um banco chinĂȘs relativos a emprĂ©stimos concedidos a pequenas e microempresas. O estudo visa desenvolver um modelo empĂ­rico que possa prever com precisĂŁo se um determinado emprĂ©stimo tem um risco aceitĂĄvel do ponto de vista do banco, ou nĂŁo. AlĂ©m disso, sĂŁo efetuados testes com diferentes mĂ©todos que permitem lidar com os problemas de classes de dados nĂŁo balanceadas e de amostras que nĂŁo refletem o problema real a modelar. Finalmente, Ă© analisada a importĂąncia relativa das variĂĄveis. O montante da dĂ­vida por pagar, a taxa de juro variĂĄvel, o prazo atĂ© a data de vencimento, a taxa de juro real, o montante do emprĂ©stimo, todas tĂȘm efeitos significativos no resultado final da previsĂŁo. O principal contributo deste estudo Ă©, assim, a construção de um modelo de avaliação de crĂ©dito que permite apoiar os bancos comerciais a identificarem com precisĂŁo o risco de crĂ©dito das pequenas e micro empresas e ajudar tambĂ©m estas empresas a superarem as suas dificuldades de crĂ©dito

    Moody’s Ratings Statistical Forecasting for Industrial and Retail Firms

    Get PDF
    Long-term ratings of companies are obtained from public data plus some additional nondisclosed information. A model based on data from firms’ public accounts is proposed to directly obtain these ratings, showing fairly close similitude with published results from Credit Rating Agencies. The rating models used to assess the creditworthiness of a firm may involve some possible conflicts of interest, as companies pay for most of the rating process and are, thus, clients of the rating firms. Such loss of faith among investors and criticism toward the rating agencies were especially severe during the financial crisis in 2008. To overcome this issue, several alternatives are addressed; in particular, the focus is on elaborating a rating model for Moody’s long-term companies’ ratings for industrial and retailing firms that could be useful as an external check of published rates. Statistical and artificial intelligence methods are used to obtain direct prediction of awarded rates in these sectors, without aggregating adjacent classes, which is usual in previous literature. This approach achieves an easy-to-replicate methodology for real rating forecasts based only on public available data, without incurring the costs associated with the rating process, while achieving a higher accuracy. With additional sampling information, these models can be extended to other sectors

    Corporate bankruptcy prediction: a comparison of logistic regression and random forest on portuguese company data

    Get PDF
    In the currentfield ofbankruptcy prediction studies, the geographical focus usually is on larger economiesrather than economies the size of Portugal. For the purpose of this studyfinancial statement data from five consecutive years prior to the event of bankruptcy in 2017 was selected. Within the data328,542healthy and unhealthy Portuguese companieswere included.Two predictive models using the Logistic Regression and Random Forest algorithm were fitted to be able to predict bankruptcy.Both developed models deliver good results even though the RandomForestmodel performs slightly better than the one based on Logistic Regression

    A Comprehensive Survey on Enterprise Financial Risk Analysis: Problems, Methods, Spotlights and Applications

    Full text link
    Enterprise financial risk analysis aims at predicting the enterprises' future financial risk.Due to the wide application, enterprise financial risk analysis has always been a core research issue in finance. Although there are already some valuable and impressive surveys on risk management, these surveys introduce approaches in a relatively isolated way and lack the recent advances in enterprise financial risk analysis. Due to the rapid expansion of the enterprise financial risk analysis, especially from the computer science and big data perspective, it is both necessary and challenging to comprehensively review the relevant studies. This survey attempts to connect and systematize the existing enterprise financial risk researches, as well as to summarize and interpret the mechanisms and the strategies of enterprise financial risk analysis in a comprehensive way, which may help readers have a better understanding of the current research status and ideas. This paper provides a systematic literature review of over 300 articles published on enterprise risk analysis modelling over a 50-year period, 1968 to 2022. We first introduce the formal definition of enterprise risk as well as the related concepts. Then, we categorized the representative works in terms of risk type and summarized the three aspects of risk analysis. Finally, we compared the analysis methods used to model the enterprise financial risk. Our goal is to clarify current cutting-edge research and its possible future directions to model enterprise risk, aiming to fully understand the mechanisms of enterprise risk communication and influence and its application on corporate governance, financial institution and government regulation

    Improving Classifier Performance Assessment of Credit Scoring Models

    Get PDF
    In evaluating credit scoring predictive power it is common to use the Re-ceiver Operating Characteristics (ROC) curve, the Area Under the Curve(AUC) and the minimum probability-weighted loss. The main weakness of the rst two assessments is not to take the costs of misclassication errors into account and the last one depends on the number of defaults in the credit portfolio. The main purposes of this paper are to provide a curve, called curve of Misclassication Error Loss (MEL), and a classier performance measure that overcome the above-mentioned drawbacks. We prove that the ROC dominance is equivalent to the MEL dominance. Furthermore, we derive the probability distribution of the proposed predictive power measure and we analyse its performance by Monte Carlo simulations. Finally, we apply the suggested methodologies to empirical data on Italian Small and Medium Enterprisers.Performance Assessment, Credit Scoring Modules, Monte Carlo simulations, Italian Enterprisers

    SME default prediction: A systematic methodology-focused review

    Get PDF
    This study reviews the methodologies used in the literature to predict failure in small and medium-sized enterprises (SMEs). We identified 145 SMEs’ default prediction studies from 1972 to early 2023. We summarized the methods used in each study. The focus points are estimation methods, sample re-balancing methods, variable selection techniques, validation methods, and variables included in the literature. More than 1,200 factors used in failure prediction models have been identified, along with 54 unique feature selection techniques and 80 unique estimation methods. Over one-third of the studies do not use any feature selection method, and more than one-quarter use only in-sample validation. Our main recommendation for researchers is to use feature selection and validate results using hold-out samples or cross-validation. As an avenue for further research, we suggest in-depth empirical comparisons of estimation methods, feature selection techniques, and sample re-balancing methods based on some large and commonly used datasets.publishedVersio

    Using neural networks and support vector machines for default prediction in South Africa

    Get PDF
    A thesis submitted to the Faculty of Computer Science and Applied Mathematics, University of Witwatersrand, in fulfillment of the requirements for the Master of Science (MSc) Johannesburg Feb 2017This is a thesis on credit risk and in particular bankruptcy prediction. It investigates the application of machine learning techniques such as support vector machines and neural networks for this purpose. This is not a thesis on support vector machines and neural networks, it simply looks at using these functions as tools to preform the analysis. Neural networks are a type of machine learning algorithm. They are nonlinear mod- els inspired from biological network of neurons found in the human central nervous system. They involve a cascade of simple nonlinear computations that when aggre- gated can implement robust and complex nonlinear functions. Neural networks can approximate most nonlinear functions, making them a quite powerful class of models. Support vector machines (SVM) are the most recent development from the machine learning community. In machine learning, support vector machines (SVMs) are su- pervised learning algorithms that analyze data and recognize patterns, used for clas- si cation and regression analysis. SVM takes a set of input data and predicts, for each given input, which of two possible classes comprises the input, making the SVM a non-probabilistic binary linear classi er. A support vector machine constructs a hyperplane or set of hyperplanes in a high or in nite dimensional space, which can be used for classi cation into the two di erent data classes. Traditional bankruptcy prediction medelling has been criticised as it makes certain underlying assumptions on the underlying data. For instance, a frequent requirement for multivarate analysis is a joint normal distribution and independence of variables. Support vector machines (and neural networks) are a useful tool for default analysis because they make far fewer assumptions on the underlying data. In this framework support vector machines are used as a classi er to discriminate defaulting and non defaulting companies in a South African context. The input data required is a set of nancial ratios constructed from the company's historic nancial statements. The data is then Divided into the two groups: a company that has defaulted and a company that is healthy (non default). The nal data sample used for this thesis consists of 23 nancial ratios from 67 companies listed on the jse. Furthermore for each company the company's probability of default is predicted. The results are benchmarked against more classical methods that are commonly used for bankruptcy prediction such as linear discriminate analysis and logistic regression. Then the results of the support vector machines, neural networks, linear discriminate analysis and logistic regression are assessed via their receiver operator curves and pro tability ratios to gure out which model is more successful at predicting default.MT 201

    Data Analytics for Credit Risk Models in Retail Banking: a new era for the banking system

    Get PDF
    Given the nature of the lending industry and its importance for global economic stability, financial institutions have always been keen on estimating the risk profile of their clients. For this reason, in the last few years several sophisticated techniques for modelling credit risk have been developed and implemented. After the financial crisis of 2007-2008, credit risk management has been further expanded and has acquired significant regulatory importance. Specifically, Basel II and III Accords have strengthened the conditions that banks must fulfil to develop their own internal models for estimating the regulatory capital and expected losses. After motivating the importance of credit risk modelling in the banking sector, in this contribution we perform a review of the traditional statistical methods used for credit risk management. Then we focus on more recent techniques based on Machine Learning techniques, and we critically compare tradition and innovation in credit risk modelling. Finally, we present a case study addressing the main steps to practically develop and validate a Probability of Default model for risk prediction via Machine Learning Techniques

    A Hybrid Technological Innovation Text Mining, Ensemble Learning and Risk Scorecard Approach for Enterprise Credit Risk Assessment

    Get PDF
    Enterprise credit risk assessment models typically use financial-based information as a predictor variable, relying on backward-looking historical information rather than forward-looking information for risk assessment. We propose a novel hybrid assessment of credit risk that uses technological innovation information as a predictor variable. Text mining techniques are used to extract this information for each enterprise. A combination of random forest and extreme gradient boosting are used for indicator screening, and finally, risk scorecard based on logistic regression is used for credit risk scoring. Our results show that technological innovation indicators obtained through text mining provide valuable information for credit risk assessment, and that the combination of ensemble learning from random forest and extreme gradient boosting combinations with logistic regression models outperforms other traditional methods. The best results achieved 0.9129 area under receiver operating characteristic. In addition, our approach provides meaningful scoring rules for credit risk assessment of technology innovation enterprises

    Improving the comparability of insolvency predictions

    Get PDF
    This working paper aims at improving the comparability of forecast quality measures of insolvency prediction studies. For this purpose, in a first step commonly used accuracy measures for categorial, ordinal and cardinal insolvency predictions are presented. It will be argued, that ordinal measures are the most suitable measures for sample spanning comparisons concerning predictive power of rating models, as they are not affected by sample default rates. A method for transforming cardinal into ordinal accuracy measures is presented, by which comparisons of insolvency prediction results of older and present-day studies are enabled. In the second part of the working paper an overview of influencing variables – aside from the quality of the insolvency prediction methods – is given, which affect the accuracy measures presented in the first part of the paper and thus impair sample spanning comparison of empirically obtained forecast quality results. In this context, methods for evaluating information losses that are attributable to the discretization of continuous rating scales or preselection of portfolios are developed. Measure results of various insolvency prognosis studies are envisaged and compared with three benchmarks. First benchmark is the accuracy that can be achieved solely by taking into account legal status and industry classification of corporations. The second benchmark is the univariate prognosis accuracy of single financial ratios. As third benchmark, ALTMAN’s Z-score model is examined, a multivariate insolvency prediction model, that is currently used as reference rating model in many empirical studies. It turns out, however, that the Z-score’s forecast quality is so discontenting, that its application is not recommendable. Instead it is suggested to use those rating models that are cited in this discussion paper, which are fully documented and which therefore can be rebuilt and directly applied to any desired data sample. If applied to the respective target groups, their performance matches with the performance of commercial rating systems, like bureau and business scores for rather small companies, middle market rating models for SMB, or agency ratings for large public companies.financial ratio analysis, corporate bankruptcy prediction, forecast validation, accuracy ratio, information entropy, sample selection, rating granularity
    • 

    corecore