12 research outputs found

    Latent deep sequential learning of behavioural sequences

    Get PDF
    The growing use of asynchronous online education (MOOCs and e-courses) in recent years has resulted in increased economic and scientific productivity, which has worsened during the coronavirus epidemic. The widespread usage of OLEs has increased enrolment, including previously excluded students, resulting in a far higher dropout rate than in conventional classrooms. Dropouts are a significant problem, especially considering the rising proliferation of online courses, from individual MOOCs to whole academic programmes due to the pandemic. Increased efficiency in dropout prevention techniques is vital for institutions, students, and faculty members and must be prioritised. In response to the resurgence of interest in the student dropout prediction (SDP) issue, there has been a significant rise in contributions to the literature on this topic. An in-depth review of the current state of the art literature on SDP is provided, with a special emphasis on Machine Learning prediction approaches; however, this is not the only focus of the thesis. We propose a complete hierarchical categorisation of the current literature that correlates to the process of design decisions in the SDP, and we demonstrate how it may be implemented. In order to enable comparative analysis, we develop a formal notation for universally defining the multiple dropout models examined by scholars in the area, including online degrees and their attributes. We look at several other important factors that have received less attention in the literature, such as evaluation metrics, acquired data, and privacy concerns. We emphasise deep sequential machine learning approaches and are considered to be one of the most successful solutions available in this field of study. Most importantly, we present a novel technique - namely GRU-AE - for tackling the SDP problem using hidden spatial information and time-related data from student trajectories. Our method is capable of dealing with data imbalances and time-series sparsity challenges. The proposed technique outperforms current methods in various situations, including the complex scenario of full-length courses (such as online degrees). This situation was thought to be less common before the outbreak, but it is now deemed important. Finally, we extend our findings to different contexts with a similar characterisation (temporal sequences of behavioural labels). Specifically, we show that our technique can be used in real-world circumstances where the unbalanced nature of the data can be mitigated by using class balancement technique (i.e. ADASYN), e.g., survival prediction in critical care telehealth systems where balancement technique alleviates the problem of inter-activity reliance and sparsity, resulting in an overall improvement in performance

    Robust Stochastic Graph Generator for Counterfactual Explanations

    Full text link
    Counterfactual Explanation (CE) techniques have garnered attention as a means to provide insights to the users engaging with AI systems. While extensively researched in domains such as medical imaging and autonomous vehicles, Graph Counterfactual Explanation (GCE) methods have been comparatively under-explored. GCEs generate a new graph similar to the original one, with a different outcome grounded on the underlying predictive model. Among these GCE techniques, those rooted in generative mechanisms have received relatively limited investigation despite demonstrating impressive accomplishments in other domains, such as artistic styles and natural language modelling. The preference for generative explainers stems from their capacity to generate counterfactual instances during inference, leveraging autonomously acquired perturbations of the input graph. Motivated by the rationales above, our study introduces RSGG-CE, a novel Robust Stochastic Graph Generator for Counterfactual Explanations able to produce counterfactual examples from the learned latent space considering a partially ordered generation sequence. Furthermore, we undertake quantitative and qualitative analyses to compare RSGG-CE's performance against SoA generative explainers, highlighting its increased ability to engendering plausible counterfactual candidates.Comment: Accepted in AAAI'2

    Adapting to Change: Robust Counterfactual Explanations in Dynamic Data Landscapes

    Full text link
    We introduce a novel semi-supervised Graph Counterfactual Explainer (GCE) methodology, Dynamic GRAph Counterfactual Explainer (DyGRACE). It leverages initial knowledge about the data distribution to search for valid counterfactuals while avoiding using information from potentially outdated decision functions in subsequent time steps. Employing two graph autoencoders (GAEs), DyGRACE learns the representation of each class in a binary classification scenario. The GAEs minimise the reconstruction error between the original graph and its learned representation during training. The method involves (i) optimising a parametric density function (implemented as a logistic regression function) to identify counterfactuals by maximising the factual autoencoder's reconstruction error, (ii) minimising the counterfactual autoencoder's error, and (iii) maximising the similarity between the factual and counterfactual graphs. This semi-supervised approach is independent of an underlying black-box oracle. A logistic regression model is trained on a set of graph pairs to learn weights that aid in finding counterfactuals. At inference, for each unseen graph, the logistic regressor identifies the best counterfactual candidate using these learned weights, while the GAEs can be iteratively updated to represent the continual adaptation of the learned graph representation over iterations. DyGRACE is quite effective and can act as a drift detector, identifying distributional drift based on differences in reconstruction errors between iterations. It avoids reliance on the oracle's predictions in successive iterations, thereby increasing the efficiency of counterfactual discovery. DyGRACE, with its capacity for contrastive learning and drift detection, will offer new avenues for semi-supervised learning and explanation generation

    A Survey on Graph Counterfactual Explanations: Definitions, Methods, Evaluation

    Full text link
    In recent years, Graph Neural Networks have reported outstanding performance in tasks like community detection, molecule classification and link prediction. However, the black-box nature of these models prevents their application in domains like health and finance, where understanding the models' decisions is essential. Counterfactual Explanations (CE) provide these understandings through examples. Moreover, the literature on CE is flourishing with novel explanation methods which are tailored to graph learning. In this survey, we analyse the existing Graph Counterfactual Explanation methods, by providing the reader with an organisation of the literature according to a uniform formal notation for definitions, datasets, and metrics, thus, simplifying potential comparisons w.r.t to the method advantages and disadvantages. We discussed seven methods and sixteen synthetic and real datasets providing details on the possible generation strategies. We highlight the most common evaluation strategies and formalise nine of the metrics used in the literature. We first introduce the evaluation framework GRETEL and how it is possible to extend and use it while providing a further dimension of comparison encompassing reproducibility aspects. Finally, we provide a discussion on how counterfactual explanation interplays with privacy and fairness, before delving into open challenges and future works.Comment: arXiv admin note: text overlap with arXiv:2107.04086 by other author

    Multimodal Motion Conditioned Diffusion Model for Skeleton-based Video Anomaly Detection

    Full text link
    Anomalies are rare and anomaly detection is often therefore framed as One-Class Classification (OCC), i.e. trained solely on normalcy. Leading OCC techniques constrain the latent representations of normal motions to limited volumes and detect as abnormal anything outside, which accounts satisfactorily for the openset'ness of anomalies. But normalcy shares the same openset'ness property, since humans can perform the same action in several ways, which the leading techniques neglect. We propose a novel generative model for video anomaly detection (VAD), which assumes that both normality and abnormality are multimodal. We consider skeletal representations and leverage state-of-the-art diffusion probabilistic models to generate multimodal future human poses. We contribute a novel conditioning on the past motion of people, and exploit the improved mode coverage capabilities of diffusion processes to generate different-but-plausible future motions. Upon the statistical aggregation of future modes, anomaly is detected when the generated set of motions is not pertinent to the actual future. We validate our model on 4 established benchmarks: UBnormal, HR-UBnormal, HR-STC, and HR-Avenue, with extensive experiments surpassing state-of-the-art results.Comment: Accepted at ICCV202

    Latent and sequential prediction of the novel coronavirus epidemiological spread

    No full text
    In this paper we present CoRoNNa a deep sequential framework for epidemic prediction that leverages a flexible combination of sequential and convolutional components to analyse the transmission of COVID-19 and, perhaps, other undiscovered viruses. Importantly, our methodology is generic and may be tailored to specific analysis goals. We exploit CoRoNNa to analyse the impact of various mobility containment policies on the pandemic using cumulative viral dissemination statistics with local demographic and movement data from several nations. Our experiments show that data on mobility has a significant, but delayed, impact on viral propagation. When compared to alternative frameworks that combine multivariate lagged predictors and basic LSTM models, CoRoNNa outperforms them. On the contrary, no technique based solely on lagged viral dissemination statistics can forecast daily cases

    CoRoNNa: A Deep Sequential Framework to Predict Epidemic Spread

    No full text
    We propose CORONNA, a deep framework for epidemic prediction to analyse the spread of COVID-19 and, potentially, of other unknown viruses, based on a flexible integration of sequential and convolutional components. Importantly, our framework is general and can be specialised according to different analysis objectives. In this paper, the specific purpose is to optimise CORONNA for analysing the impact of different mobility containment policies on the epidemic. To this end, we integrate cumulative viral diffusion statistics and local demographic and mobility information of several countries. Our analysis confirms that mobility data have a strong, but delayed, effect on the viral spread. In this context, CORONNA has superior performances when compared with other frameworks that incorporate multivariate lagged predictors, and with simple LSTM models. On the contrary, no method is able to predict daily cases based only on lagged viral diffusion statistics

    Towards Non-Adversarial Algorithmic Recourse

    Full text link
    The streams of research on adversarial examples and counterfactual explanations have largely been growing independently. This has led to several recent works trying to elucidate their similarities and differences. Most prominently, it has been argued that adversarial examples, as opposed to counterfactual explanations, have a unique characteristic in that they lead to a misclassification compared to the ground truth. However, the computational goals and methodologies employed in existing counterfactual explanation and adversarial example generation methods often lack alignment with this requirement. Using formal definitions of adversarial examples and counterfactual explanations, we introduce non-adversarial algorithmic recourse and outline why in high-stakes situations, it is imperative to obtain counterfactual explanations that do not exhibit adversarial characteristics. We subsequently investigate how different components in the objective functions, e.g., the machine learning model or cost function used to measure distance, determine whether the outcome can be considered an adversarial example or not. Our experiments on common datasets highlight that these design choices are often more critical in deciding whether recourse is non-adversarial than whether recourse or attack algorithms are used. Furthermore, we show that choosing a robust and accurate machine learning model results in less adversarial recourse desired in practice

    Hidden space deep sequential risk prediction on student trajectories

    No full text
    Online learning environments (OLEs) have seen a continuous increase over the past decade and a sudden surge in the last year, due to the coronavirus outbreak. The widespread use of OLEs has led to an increasing number of enrolments, even from students who had previously left education systems, but it also resulted in a much higher dropout rate than in traditional classrooms. This is a crucial problem since online courses have rapidly expanded from individual MOOCs to entire study programmes, also due to the pandemic. Early detection of students in difficulty is a challenging problem that can be mitigated with the support of state-of-the-art methods for data analytics and machine learning. In this study, we propose a novel strategy that exploits both hidden space information and time-related data from student trajectories and copes with unbalanced data and time-series sparsity issues to solve the student dropout prediction problem. The proposed approach outperforms state-of-the-art methods, particularly in the complex case of full-length curricula (such as online degrees), a scenario that was thought to be less common before the pandemic, but is now particularly relevant

    House in the (biometric) cloud: a possible application

    No full text
    This article presents a novel approach to extend cloud computing from company services to consumer biometrics. The proposed system recognizes the person at the door, allowing entrance or denying it according to the recognition result. Very little processing is required locally, and biometrics is implemented as a service
    corecore