1,485 research outputs found

    Graph based Anomaly Detection and Description: A Survey

    Get PDF
    Detecting anomalies in data is a vital task, with numerous high-impact applications in areas such as security, finance, health care, and law enforcement. While numerous techniques have been developed in past years for spotting outliers and anomalies in unstructured collections of multi-dimensional points, with graph data becoming ubiquitous, techniques for structured graph data have been of focus recently. As objects in graphs have long-range correlations, a suite of novel technology has been developed for anomaly detection in graph data. This survey aims to provide a general, comprehensive, and structured overview of the state-of-the-art methods for anomaly detection in data represented as graphs. As a key contribution, we give a general framework for the algorithms categorized under various settings: unsupervised vs. (semi-)supervised approaches, for static vs. dynamic graphs, for attributed vs. plain graphs. We highlight the effectiveness, scalability, generality, and robustness aspects of the methods. What is more, we stress the importance of anomaly attribution and highlight the major techniques that facilitate digging out the root cause, or the ‘why’, of the detected anomalies for further analysis and sense-making. Finally, we present several real-world applications of graph-based anomaly detection in diverse domains, including financial, auction, computer traffic, and social networks. We conclude our survey with a discussion on open theoretical and practical challenges in the field

    A simulation-driven approach to non-compliance

    Get PDF
    This dissertation proposes a methodological framework for the use of simulation-based methods to investigate questions of non-compliance in a legal context. Its aim is to generate observed or previously unobserved instances of non-compliance and use them to improve compliance and trust in a given socio-economic infrastructure. The framework consists of three components: a normative system implemented as an agent-based model, a profit-driven agent generating instances of non-compliance, and a formalization process transforming the generated behavior into a formal model.The most sophisticated ways of law-breaking are typically associated with economic crime. For this reason, we investigated three case studies in the financial domain. The first case study develops an agent-based model investigating the collective response of compliant agents to market disturbances originated by fraudulent activity, as during the U.S. subprime mortgage crisis in 2007. The second case study investigates the price evolution in the Bitcoin market under the influence of the price manipulation that occurred in 2017/18. The third case study investigates Ponzi schemes on smart contracts. All case studies showed a high level of agreement with qualitative and quantitative observations. Identification, extraction, and formalization of non-compliant behavior generated via simulation is a central topic in the later chapters of the thesis. We introduce a method that considers fraudulent schemes as neighborhoods of profitable non-compliant behavior. We illustrate the method on a grid environment with a path-finding agent. This simplified case study has been chosen as it captures fundamental features of non-compliance, yet, further generalization is needed for real-world scenarios

    A simulation-driven approach to non-compliance

    Get PDF
    This dissertation proposes a methodological framework for the use of simulation-based methods to investigate questions of non-compliance in a legal context. Its aim is to generate observed or previously unobserved instances of non-compliance and use them to improve compliance and trust in a given socio-economic infrastructure. The framework consists of three components: a normative system implemented as an agent-based model, a profit-driven agent generating instances of non-compliance, and a formalization process transforming the generated behavior into a formal model.The most sophisticated ways of law-breaking are typically associated with economic crime. For this reason, we investigated three case studies in the financial domain. The first case study develops an agent-based model investigating the collective response of compliant agents to market disturbances originated by fraudulent activity, as during the U.S. subprime mortgage crisis in 2007. The second case study investigates the price evolution in the Bitcoin market under the influence of the price manipulation that occurred in 2017/18. The third case study investigates Ponzi schemes on smart contracts. All case studies showed a high level of agreement with qualitative and quantitative observations. Identification, extraction, and formalization of non-compliant behavior generated via simulation is a central topic in the later chapters of the thesis. We introduce a method that considers fraudulent schemes as neighborhoods of profitable non-compliant behavior. We illustrate the method on a grid environment with a path-finding agent. This simplified case study has been chosen as it captures fundamental features of non-compliance, yet, further generalization is needed for real-world scenarios

    Surgical Management of Gastroesophageal Reflux in Children: Risk Stratification and Prediction of Outcomes

    Get PDF
    Introduction: Since the 1980s fundoplication, an operation developed for adults with hiatus hernia and reflux symptoms, has been performed in children with gastroesophageal reflux disease (GORD). When compared to adult outcomes, paediatric fundoplication has resulted in higher failure and revision rates. In the first chapter we explore differences in paradigm, patient population and outcomes. Firstly, symptoms are poorly defined and are measured by instruments of varying quality. Secondly, neurological impairment (NI), prematurity and congenital anomalies (oesophageal atresia, congenital diaphragmatic hernia) are prevalent in children. / Purpose: To develop methods for stratifying paediatric fundoplication risk and predicting outcomes based on symptom profile, demographic factors, congenital and medical history. / Methods: Study objectives are addressed in three opera: a symptom questionnaire development (TARDIS:REFLUX), a randomised controlled trial (RCT) and a retrospective database study (RDS). TARDIS: REFLUX: In the second chapter, digital research methods are used to design and validate a symptom questionnaire for paediatric GORD. The questionnaire is a market-viable smartphone app hosted on a commercial platform and trialed in a clinical pilot study. / RCT: In the third chapter, the REMOS trial is reported. The trial addresses the subset of children with NI and feeding difficulties. Participants are randomized to gastrostomy with or without fundoplication. Notably, pre- and post-operative reflux is quantified using pH-impedance. / RDS: In the fourth chapter, data mining and machine learning strategies are applied to a retrospective paediatric GORD database. Predictive modelling techniques applied include logistic regression, decision trees, random forests and market basket analysis. / Results and conclusion: This work makes two key contributions. Firstly, an effective methodology for development of digital research tools is presented here. Secondly, a synthesis is made of literature, the randomised controlled trial and retrospective database modelling. The resulting product is an evidence-based algorithm for the surgical management of children with GORD

    Pragmatism and methodology: doing research that matters with mixed methods

    Get PDF
    DescriptionContentsResourcesCoursesAbout the Authors Taking a pragmatist approach to methods and methodology that fosters meaningful, impactful, and ethical research, this book rises to the challenge of today's data revolution. It shows how pragmatism can turn challenges, such as the abundance and accumulation of big qualitative data, into opportunities. The authors summarize the pragmatist approach to different aspects of research, from epistemology, theory, and questions to ethics, as well as data collection and analysis. The chapters outline and document a new type of mixed methods design called 'multi-resolution research,” which serves to overcome old divides between quantitative and qualitative methods. It is the ideal resource for students and researchers within the social and behavioural sciences seeking new ways to analyze large sets of qualitative data. This book is also available as Open Access on Cambridge Core

    Locomotion Traces Data Mining for Supporting Frail People with Cognitive Impairment

    Get PDF
    The rapid increase in the senior population is posing serious challenges to national healthcare systems. Hence, innovative tools are needed to early detect health issues, including cognitive decline. Several clinical studies show that it is possible to identify cognitive impairment based on the locomotion patterns of older people. Thus, this thesis at first focused on providing a systematic literature review of locomotion data mining systems for supporting Neuro-Degenerative Diseases (NDD) diagnosis, identifying locomotion anomaly indicators and movement patterns for discovering low-level locomotion indicators, sensor data acquisition, and processing methods, as well as NDD detection algorithms considering their pros and cons. Then, we investigated the use of sensor data and Deep Learning (DL) to recognize abnormal movement patterns in instrumented smart-homes. In order to get rid of the noise introduced by indoor constraints and activity execution, we introduced novel visual feature extraction methods for locomotion data. Our solutions rely on locomotion traces segmentation, image-based extraction of salient features from locomotion segments, and vision-based DL. Furthermore, we proposed a data augmentation strategy to increase the volume of collected data and generalize the solution to different smart-homes with different layouts. We carried out extensive experiments with a large real-world dataset acquired in a smart-home test-bed from older people, including people with cognitive diseases. Experimental comparisons show that our system outperforms state-of-the-art methods

    Data quality assurance for strategic decision making in Abu Dhabi's public organisations

    Get PDF
    “A thesis submitted to the University of Bedfordshire, in partial fulfilment of the requirements for the degree of Master of Philosophy”.Data quality is an important aspect of an organisation’s strategies for supporting decision makers in reaching the best decisions possible and consequently attaining the organisation’s objectives. In the case of public organisations, decisions ultimately concern the public and hence further diligence is required to make sure that these decisions do, for instance, preserve economic resources, maintain public health, and provide national security. The decision making process requires a wealth of information in order to achieve efficient results. Public organisations typically acquire great amounts of data generated by public services. However, the vast amount of data stored in public organisations’ databases may be one of the main reasons for inefficient decisions made by public organisations. Processing vast amounts of data and extracting accurate information are not easy tasks. Although technology helps in this respect, for example, the use of decision support systems, it is not sufficient for improving decisions to a significant level of assurance. The research proposed using data mining to improve results obtained by decision support systems. However, more considerations are needed than the mere technological aspects. The research argues that a complete data quality framework is needed in order to improve data quality and consequently the decision making process in public organisations. A series of surveys conducted in seven public organisations in Abu Dhabi Emirate of the United Arab Emirates contributed to the design of a data quality framework. The framework comprises elements found necessary to attain the quality of data reaching decision makers. The framework comprises seven elements ranging from technical to human-based found important to attain data quality in public organisations taking Abu Dhabi public organisations as the case. The interaction and integration of these elements contributes to the quality of data reaching decision makers and hence to the efficiency of decisions made by public organisations. The framework suggests that public organisations may need to adopt a methodological basis to support the decision making process. This includes more training courses and supportive bodies of the organisational units, such as decision support centres, information security and strategic management. The framework also underscores the importance of acknowledging human and cultural factors involved in the decision making process. Such factors have implications for how training and raising awareness are implemented to lead to effective methods of system development
    • 

    corecore