1,984 research outputs found

    Un enfoque de sustentabilidad utilizando lĂłgica difusa y minerĂ­a de datos

    Get PDF
    [ES] Sustainable development goals are now the agreed criteria to monitor states, and this work will demonstrate that numerical and graphical methods are valuable tools in assessing progress. Fuzzy Logic is a reliable procedure for transforming human qualitative knowledge into quantitative variables that can be used in the reasoning of the type “if, then” to obtain answers pertaining to sustainability assessment. Applications of machine learning techniques and artificial intelligence procedures span almost all fields of science. Here, for the first-time, unsupervised machine learning is applied to sustainability assessment, combining numerical approaches with graphical procedures to analyze global sustainability. CD HJ-Biplots to portray graphically the sustainability position of a large number of countries are a useful complement to mathematical models of sustainability. Graphical information could be useful to planners it shows directly how countries are grouped according to the most related sustainability indicators. Thus, planners can prioritize social, environmental, and economic policies and make the most effective decisions. One could graphically observe the dynamic evolution of sustainability worldwide over time with a graphical approach used to draw relevant conclusions. In an era of climate change, species extinction, poverty, and environmental migration, such observations could aid political decision-making regarding the future of our planet. A large number of countries remain in the areas of moderate or low sustainability. Fuzzy logic has proven to be an uncontested numerical method as it occurs with SAFE. An unsupervised learning method called Variational Autoencoder interplay Graphical Analysis (VEA&GA) has been proposed, to support sustainability performance with appropriate training data. The promising results show that this can be a sound alternative to assess sustainability, extrapolating its applications to other kinds of problems at different levels of analysis (continents, regions, cities, etc.) further corroborating the effectiveness of the unsupervised training methods

    Unsupervised learning for anomaly detection in Australian medical payment data

    Full text link
    Fraudulent or wasteful medical insurance claims made by health care providers are costly for insurers. Typically, OECD healthcare organisations lose 3-8% of total expenditure due to fraud. As Australia’s universal public health insurer, Medicare Australia, spends approximately A34billionperannumontheMedicareBenefitsSchedule(MBS)andPharmaceuticalBenefitsScheme,wastedspendingofA 34 billion per annum on the Medicare Benefits Schedule (MBS) and Pharmaceutical Benefits Scheme, wasted spending of A1–2.7 billion could be expected.However, fewer than 1% of claims to Medicare Australia are detected as fraudulent, below international benchmarks. Variation is common in medicine, and health conditions, along with their presentation and treatment, are heterogenous by nature. Increasing volumes of data and rapidly changing patterns bring challenges which require novel solutions. Machine learning and data mining are becoming commonplace in this field, but no gold standard is yet available. In this project, requirements are developed for real-world application to compliance analytics at the Australian Government Department of Health and Aged Care (DoH), covering: unsupervised learning; problem generalisation; human interpretability; context discovery; and cost prediction. Three novel methods are presented which rank providers by potentially recoverable costs. These methods used association analysis, topic modelling, and sequential pattern mining to provide interpretable, expert-editable models of typical provider claims. Anomalous providers are identified through comparison to the typical models, using metrics based on costs of excess or upgraded services. Domain knowledge is incorporated in a machine-friendly way in two of the methods through the use of the MBS as an ontology. Validation by subject-matter experts and comparison to existing techniques shows that the methods perform well. The methods are implemented in a software framework which enables rapid prototyping and quality assurance. The code is implemented at the DoH, and further applications as decision-support systems are in progress. The developed requirements will apply to future work in this fiel

    In-Network Outlier Detection in Wireless Sensor Networks

    Full text link
    To address the problem of unsupervised outlier detection in wireless sensor networks, we develop an approach that (1) is flexible with respect to the outlier definition, (2) computes the result in-network to reduce both bandwidth and energy usage,(3) only uses single hop communication thus permitting very simple node failure detection and message reliability assurance mechanisms (e.g., carrier-sense), and (4) seamlessly accommodates dynamic updates to data. We examine performance using simulation with real sensor data streams. Our results demonstrate that our approach is accurate and imposes a reasonable communication load and level of power consumption.Comment: Extended version of a paper appearing in the Int'l Conference on Distributed Computing Systems 200

    Latent tree models

    Full text link
    Latent tree models are graphical models defined on trees, in which only a subset of variables is observed. They were first discussed by Judea Pearl as tree-decomposable distributions to generalise star-decomposable distributions such as the latent class model. Latent tree models, or their submodels, are widely used in: phylogenetic analysis, network tomography, computer vision, causal modeling, and data clustering. They also contain other well-known classes of models like hidden Markov models, Brownian motion tree model, the Ising model on a tree, and many popular models used in phylogenetics. This article offers a concise introduction to the theory of latent tree models. We emphasise the role of tree metrics in the structural description of this model class, in designing learning algorithms, and in understanding fundamental limits of what and when can be learned

    Machine learning methods for generating high dimensional discrete datasets

    Get PDF
    The development of platforms and techniques for emerging Big Data and Machine Learning applications requires the availability of real-life datasets. A possible solution is to synthesize datasets that reflect patterns of real ones using a two-step approach: first, a real dataset X is analyzed to derive relevant patterns Z and, then, to use such patterns for reconstructing a new dataset X ' that preserves the main characteristics of X. This survey explores two possible approaches: (1) Constraint-based generation and (2) probabilistic generative modeling. The former is devised using inverse mining (IFM) techniques, and consists of generating a dataset satisfying given support constraints on the itemsets of an input set, that are typically the frequent ones. By contrast, for the latter approach, recent developments in probabilistic generative modeling (PGM) are explored that model the generation as a sampling process from a parametric distribution, typically encoded as neural network. The two approaches are compared by providing an overview of their instantiations for the case of discrete data and discussing their pros and cons. This article is categorized under: Fundamental Concepts of Data and Knowledge > Big Data Mining Technologies > Machine Learning Algorithmic Development > Structure Discover

    Machine Learning Methods for Generating High Dimensional Discrete Datasets

    Get PDF
    The development of platforms and techniques for emerging Big Data and Machine Learning applications requires the availability of real-life datasets. A possible solution is to synthesize datasets that reflect patterns of real ones using a two-step approach: first, a real dataset X is analyzed to derive relevant patterns Z and, then, to use such patterns for reconstructing a new dataset X\u27 that preserves the main characteristics of X. This survey explores two possible approaches: (1) Constraint-based generation and (2) probabilistic generative modeling. The former is devised using inverse mining (IFM) techniques, and consists of generating a dataset satisfying given support constraints on the itemsets of an input set, that are typically the frequent ones. By contrast, for the latter approach, recent developments in probabilistic generative modeling (PGM) are explored that model the generation as a sampling process from a parametric distribution, typically encoded as neural network. The two approaches are compared by providing an overview of their instantiations for the case of discrete data and discussing their pros and cons

    Temporal and spatial trends of adult mortality in small areas of Brazil, 1980–2010

    Get PDF
    To determine the variations and spatial patterns of adult mortality across regions, over time, and by sex for 137 small areas in Brazil, we first apply TOPALS to estimate and smooth mortality rates and then use death distribution methods to evaluate the quality of the mortality data. Lastly, we employ spatial autocorrelation statistics and cluster analysis to identify the adult mortality trends and variations in these areas between 1980 and 2010. We find not only that regions in Brazil’s South and Southeast already had complete death registration systems prior to the study period, but that the completeness of death count coverage improved over time across the entire nation—most especially in lesser developed regions—probably because of public investment in health data collection. By also comparing adult mortality by sex and by region, we document a mortality sex differential in favor of women that remains high over the entire study period, most probably as a result of increased morbidity from external causes, especially among males. This increase also explains the concentration of high male mortality levels in some areas76CONSELHO NACIONAL DE DESENVOLVIMENTO CIENTÍFICO E TECNOLÓGICO - CNPQCOORDENAÇÃO DE APERFEIÇOAMENTO DE PESSOAL DE NÍVEL SUPERIOR - CAPES470866/2014-4; 454223/2014-5não te

    Artificial Intelligence for Sustainability—A Systematic Review of Information Systems Literature

    Get PDF
    The booming adoption of Artificial Intelligence (AI) likewise poses benefits and challenges. In this paper, we particularly focus on the bright side of AI and its promising potential to face our society’s grand challenges. Given this potential, different studies have already conducted valuable work by conceptualizing specific facets of AI and sustainability, including reviews on AI and Information Systems (IS) research or AI and business values. Nonetheless, there is still little holistic knowledge at the intersection of IS, AI, and sustainability. This is problematic because the IS discipline, with its socio-technical nature, has the ability to integrate perspectives beyond the currently dominant technological one as well as can advance both theory and the development of purposeful artifacts. To bridge this gap, we disclose how IS research currently makes use of AI to boost sustainable development. Based on a systematically collected corpus of 95 articles, we examine sustainability goals, data inputs, technologies and algorithms, and evaluation approaches that coin the current state of the art within the IS discipline. This comprehensive overview enables us to make more informed investments (e.g., policy and practice) as well as to discuss blind spots and possible directions for future research
    • 

    corecore