18 research outputs found

    Naive Bayesian Classification of Structured Data

    Full text link

    DATA MINING CLASSIFICATION ALGORITHMS FOR DIABETES DATASET USING WEKA TOOL

    Get PDF
    Data mining explores a huge amount of data to extract the information to be meaningful. In the field of public health, data mining hold a crucial contribution in predicting disease in early stage. In order to detect diseases, the patients need to conduct various tests. In the context of disease predicion, Data mining techniques aims to reduce the test that patients need to accomplish. Also the techniques is used to increase the accuracy rate of detection. Nowadays, diabetes attacks many adults in the world. Moreover, in order to reduce the number of adult having diabetes, an effective and efficient diabetes detection mechanism should be found. This report will apply some data mining techniques on diabetes dataset that has been downloaded at UCI Machine Learning Repository.Three kind of classification algorithm such as Naïve Bayes Classifier, Multilayer Perceptrons (MLP’s) and Desicion Tree (J.48) have been performed on this dataset. Obtained outcomes indicated that Naïve Bayes Classifier achieved the highest accuracy with 76,30%. As the result, this algorithm is a good method to classify and diagnose diabetes diseases on studying dataset

    Wallenius Naive Bayes

    Get PDF
    Traditional event models underlying naive Bayes classifiers assume probability distributions that are not appropriate for binary data generated by human behaviour. In this work, we develop a new event model, based on a somewhat forgotten distribution created by Kenneth Ted Wallenius in 1963. We show that it achieves superior performance using less data on a collection of Facebook datasets, where the task is to predict personality traits, based on likes.Faculty of Applied Economics, University of Antwerp, Belgium; Department of Information, Operations & Management Sciences, NYU Stern School of Busines

    Ensemble Learning based Anomaly Detection for IoT Cybersecurity via Bayesian Hyperparameters Sensitivity Analysis

    Full text link
    The Internet of Things (IoT) integrates more than billions of intelligent devices over the globe with the capability of communicating with other connected devices with little to no human intervention. IoT enables data aggregation and analysis on a large scale to improve life quality in many domains. In particular, data collected by IoT contain a tremendous amount of information for anomaly detection. The heterogeneous nature of IoT is both a challenge and an opportunity for cybersecurity. Traditional approaches in cybersecurity monitoring often require different kinds of data pre-processing and handling for various data types, which might be problematic for datasets that contain heterogeneous features. However, heterogeneous types of network devices can often capture a more diverse set of signals than a single type of device readings, which is particularly useful for anomaly detection. In this paper, we present a comprehensive study on using ensemble machine learning methods for enhancing IoT cybersecurity via anomaly detection. Rather than using one single machine learning model, ensemble learning combines the predictive power from multiple models, enhancing their predictive accuracy in heterogeneous datasets rather than using one single machine learning model. We propose a unified framework with ensemble learning that utilises Bayesian hyperparameter optimisation to adapt to a network environment that contains multiple IoT sensor readings. Experimentally, we illustrate their high predictive power when compared to traditional methods

    Modeling Complex Networks For (Electronic) Commerce

    Get PDF
    NYU, Stern School of Business, IOMS Department, Center for Digital Economy Researc

    Artificial Intelligence in Materials Modeling and Design

    Get PDF
    In recent decades, the use of artificial intelligence (AI) techniques in the field of materials modeling has received significant attention owing to their excellent ability to analyze a vast amount of data and reveal correlations between several complex interrelated phenomena. In this review paper, we summarize recent advances in the applications of AI techniques for numerical modeling of different types of materials. AI techniques such as machine learning and deep learning show great advantages and potential for predicting important mechanical properties of materials and reveal how changes in certain principal parameters affect the overall behavior of engineering materials. Furthermore, in this review, we show that the application of AI techniques can significantly help to improve the design and optimize the properties of future advanced engineering materials. Finally, a perspective on the challenges and prospects of the applications of AI techniques for material modeling is presented

    Spell checker implementation to analyze the narrative essay of sixth-grade elementary school students in Indonesia

    Get PDF
    Learning Bahasa Indonesia correctly in terms of writing and reading is needed by students to understand well the language learned in school. Some problems arise when elementary school students still need guidance in writing Indonesian sentences in narratives, that are still lacking the standard or misspelling. Teachers usually read essays from students but it takes a lot for the teacher to learn. This research emerged with the aim of assisting teachers and students in correcting the spelling of the essay that was written so that it would be a perfect and perfect sentence. Applications built using internet technology so students can access the system anywhere. The results show that 87% of students experience spelling errors in writing narrative essay because they are inaccurate to what they write, so the application will provide automatic correction and already implemented in several schools at Malang

    Wallenius Naive Bayes

    Get PDF
    Traditional event models underlying naive Bayes classifiers assume probability distributions that are not appropriate for binary data generated by human behaviour. In this work, we develop a new event model, based on a somewhat forgotten distribution created by Kenneth Ted Wallenius in 1963. We show that it achieves superior performance using less data on a collection of Facebook datasets, where the task is to predict personality traits, based on likes.Faculty of Applied Economics, University of Antwerp, Belgium; Department of Information, Operations & Management Sciences, NYU Stern School of Busines

    Probabilistic inductive constraint logic

    Get PDF
    AbstractProbabilistic logical models deal effectively with uncertain relations and entities typical of many real world domains. In the field of probabilistic logic programming usually the aim is to learn these kinds of models to predict specific atoms or predicates of the domain, called target atoms/predicates. However, it might also be useful to learn classifiers for interpretations as a whole: to this end, we consider the models produced by the inductive constraint logic system, represented by sets of integrity constraints, and we propose a probabilistic version of them. Each integrity constraint is annotated with a probability, and the resulting probabilistic logical constraint model assigns a probability of being positive to interpretations. To learn both the structure and the parameters of such probabilistic models we propose the system PASCAL for "probabilistic inductive constraint logic". Parameter learning can be performed using gradient descent or L-BFGS. PASCAL has been tested on 11 datasets and compared with a few statistical relational systems and a system that builds relational decision trees (TILDE): we demonstrate that this system achieves better or comparable results in terms of area under the precision–recall and receiver operating characteristic curves, in a comparable execution time
    corecore