39,203 research outputs found

    Hybrid model using logit and nonparametric methods for predicting micro-entity failure

    Get PDF
    Following the calls from literature on bankruptcy, a parsimonious hybrid bankruptcy model is developed in this paper by combining parametric and non-parametric approaches.To this end, the variables with the highest predictive power to detect bankruptcy are selected using logistic regression (LR). Subsequently, alternative non-parametric methods (Multilayer Perceptron, Rough Set, and Classification-Regression Trees) are applied, in turn, to firms classified as either “bankrupt” or “not bankrupt”. Our findings show that hybrid models, particularly those combining LR and Multilayer Perceptron, offer better accuracy performance and interpretability and converge faster than each method implemented in isolation. Moreover, the authors demonstrate that the introduction of non-financial and macroeconomic variables complement financial ratios for bankruptcy prediction

    A methodology for the selection of new technologies in the aviation industry

    Get PDF
    The purpose of this report is to present a technology selection methodology to quantify both tangible and intangible benefits of certain technology alternatives within a fuzzy environment. Specifically, it describes an application of the theory of fuzzy sets to hierarchical structural analysis and economic evaluations for utilisation in the industry. The report proposes a complete methodology to accurately select new technologies. A computer based prototype model has been developed to handle the more complex fuzzy calculations. Decision-makers are only required to express their opinions on comparative importance of various factors in linguistic terms rather than exact numerical values. These linguistic variable scales, such as ‘very high’, ‘high’, ‘medium’, ‘low’ and ‘very low’, are then converted into fuzzy numbers, since it becomes more meaningful to quantify a subjective measurement into a range rather than in an exact value. By aggregating the hierarchy, the preferential weight of each alternative technology is found, which is called fuzzy appropriate index. The fuzzy appropriate indices of different technologies are then ranked and preferential ranking orders of technologies are found. From the economic evaluation perspective, a fuzzy cash flow analysis is employed. This deals quantitatively with imprecision or uncertainties, as the cash flows are modelled as triangular fuzzy numbers which represent ‘the most likely possible value’, ‘the most pessimistic value’ and ‘the most optimistic value’. By using this methodology, the ambiguities involved in the assessment data can be effectively represented and processed to assure a more convincing and effective decision- making process when selecting new technologies in which to invest. The prototype model was validated with a case study within the aviation industry that ensured it was properly configured to meet the

    Qualitative Effects of Knowledge Rules in Probabilistic Data Integration

    Get PDF
    One of the problems in data integration is data overlap: the fact that different data sources have data on the same real world entities. Much development time in data integration projects is devoted to entity resolution. Often advanced similarity measurement techniques are used to remove semantic duplicates from the integration result or solve other semantic conflicts, but it proofs impossible to get rid of all semantic problems in data integration. An often-used rule of thumb states that about 90% of the development effort is devoted to solving the remaining 10% hard cases. In an attempt to significantly decrease human effort at data integration time, we have proposed an approach that stores any remaining semantic uncertainty and conflicts in a probabilistic database enabling it to already be meaningfully used. The main development effort in our approach is devoted to defining and tuning knowledge rules and thresholds. Rules and thresholds directly impact the size and quality of the integration result. We measure integration quality indirectly by measuring the quality of answers to queries on the integrated data set in an information retrieval-like way. The main contribution of this report is an experimental investigation of the effects and sensitivity of rule definition and threshold tuning on the integration quality. This proves that our approach indeed reduces development effort — and not merely shifts the effort to rule definition and threshold tuning — by showing that setting rough safe thresholds and defining only a few rules suffices to produce a ‘good enough’ integration that can be meaningfully used
    corecore