6,463 research outputs found

    A distributed framework for semi-automatically developing architectures of brain and mind

    Get PDF
    Developing comprehensive theories of low-level neuronal brain processes and high-level cognitive behaviours, as well as integrating them, is an ambitious challenge that requires new conceptual, computational, and empirical tools. Given the complexities of these theories, they will almost certainly be expressed as computational systems. Here, we propose to use recent developments in grid technology to develop a system of evolutionary scientific discovery, which will (a) enable empirical researchers to make their data widely available for use in developing and testing theories, and (b) enable theorists to semi-automatically develop computational theories. We illustrate these ideas with a case study taken from the domain of categorisation

    On Cognitive Preferences and the Plausibility of Rule-based Models

    Get PDF
    It is conventional wisdom in machine learning and data mining that logical models such as rule sets are more interpretable than other models, and that among such rule-based models, simpler models are more interpretable than more complex ones. In this position paper, we question this latter assumption by focusing on one particular aspect of interpretability, namely the plausibility of models. Roughly speaking, we equate the plausibility of a model with the likeliness that a user accepts it as an explanation for a prediction. In particular, we argue that, all other things being equal, longer explanations may be more convincing than shorter ones, and that the predominant bias for shorter models, which is typically necessary for learning powerful discriminative models, may not be suitable when it comes to user acceptance of the learned models. To that end, we first recapitulate evidence for and against this postulate, and then report the results of an evaluation in a crowd-sourcing study based on about 3.000 judgments. The results do not reveal a strong preference for simple rules, whereas we can observe a weak preference for longer rules in some domains. We then relate these results to well-known cognitive biases such as the conjunction fallacy, the representative heuristic, or the recogition heuristic, and investigate their relation to rule length and plausibility.Comment: V4: Another rewrite of section on interpretability to clarify focus on plausibility and relation to interpretability, comprehensibility, and justifiabilit

    A survey on utilization of data mining approaches for dermatological (skin) diseases prediction

    Get PDF
    Due to recent technology advances, large volumes of medical data is obtained. These data contain valuable information. Therefore data mining techniques can be used to extract useful patterns. This paper is intended to introduce data mining and its various techniques and a survey of the available literature on medical data mining. We emphasize mainly on the application of data mining on skin diseases. A categorization has been provided based on the different data mining techniques. The utility of the various data mining methodologies is highlighted. Generally association mining is suitable for extracting rules. It has been used especially in cancer diagnosis. Classification is a robust method in medical mining. In this paper, we have summarized the different uses of classification in dermatology. It is one of the most important methods for diagnosis of erythemato-squamous diseases. There are different methods like Neural Networks, Genetic Algorithms and fuzzy classifiaction in this topic. Clustering is a useful method in medical images mining. The purpose of clustering techniques is to find a structure for the given data by finding similarities between data according to data characteristics. Clustering has some applications in dermatology. Besides introducing different mining methods, we have investigated some challenges which exist in mining skin data

    Data Engineering for the Analysis of Semiconductor Manufacturing Data

    Get PDF
    We have analyzed manufacturing data from several different semiconductor manufacturing plants, using decision tree induction software called Q-YIELD. The software generates rules for predicting when a given product should be rejected. The rules are intended to help the process engineers improve the yield of the product, by helping them to discover the causes of rejection. Experience with Q-YIELD has taught us the importance of data engineering -- preprocessing the data to enable or facilitate decision tree induction. This paper discusses some of the data engineering problems we have encountered with semiconductor manufacturing data. The paper deals with two broad classes of problems: engineering the features in a feature vector representation and engineering the definition of the target concept (the classes). Manufacturing process data present special problems for feature engineering, since the data have multiple levels of granularity (detail, resolution). Engineering the target concept is important, due to our focus on understanding the past, as opposed to the more common focus in machine learning on predicting the future

    Advanced Analytics Success Factors - A Case Study

    Get PDF
    Companies are increasingly taking into use advanced analytics solutions. Advanced analytics solutions are computer programs that analyze data, make predictions on the future, and give optimization-based recommendations on courses of action for achieving pre-determined business goals. Analytics solutions employ sophisticated statistical and mathematical models, and are often offered by third parties. Companies use analytics solutions to improve the efficiency of their operations. This thesis studies whether the distinction between analytics and advanced analytics made in literature is well-founded. The second aim of this study is to find out, what contributes to an analytics initiative’s success. The study begins with a literature review synthesizing the findings of previous analytics research. The resulting synthesis identifies four distinct stages in an analytics project. They are acquiring data, transforming it into insights, communicating the insights, making business decisions, and finally implementing the decisions. Factors that contribute to each stage’s success are identified. The hypotheses that were developed in the theoretical part of the thesis are subsequently tested empirically using the single case study method and semi-structured interviews. The case study confirms the findings of earlier research. Analytics can be viewed as a process with clearly identifiable stages. Specific measures can be taken to improve the success of each stage. The results obtained suggest that an analytics initiative should always be preceded by a thorough goal definition stage. This is a finding that earlier research has not emphasized sufficiently. The study offers business executives a clear roadmap for managing analytics initiatives. It formulates clear action points and allocates parties the responsibility for executing them. The study also highlights some ordinary pitfalls preventing companies from fully benefitting from the results of analytics initiatives. Finally, the study points out new interesting research opportunities in the intersection of information systems science and cognitive science. A key difficulty in using analytics effectively is that the reasoning behind the insights created by the solutions are often complex. Cognitive science could provide us tools for making the insights easier to digest. Lastly, the study highlights that process decoupling will eventually be applied to analytics initiatives. Future studies should research how the stages of an analytics initiative can be separated from each other, and outsourced to parties performing them the most effectively
    • …
    corecore