59 research outputs found

    UGURU: a natural language UNIX consultant

    Get PDF
    UGURU is a natural language conversation program, implemented in Prolog, which can manage a wide knowledge base of facts about Unix. The range and wording of questions that it understands are based on surveys taken of students, mostly Unix beginners. UGURU is also designed to accept statements in English that can be added as facts to the knowledge base. Each fact is represented as a binding set: a verb-oriented semantic net with the characteristics of directed acyclic graphs. The main actions taken by UGURU are divided between two primary modules, a parser and a retriever. To produce a binding set from an input, the parser incorporates a new kind of object-oriented grammar of several levels, parallel tracing of distinct parse trees by independent units called recognizers, the concurrent use of both syntactic and semantic knowledge, and a pragmatic criterion that requires the system to mimic the sequence of human parsing. The retriever, invoked to answer input questions, seeks to match the binding set representing the question to a fact in the knowledge base by performing semantic transformations on the two sets

    Adversarial Data Augmentation for HMM-based Anomaly Detection

    Get PDF
    In this work, we concentrate on the detection of anomalous behaviors in systems operating in the physical world and for which it is usually not possible to have a complete set of all possible anomalies in advance. We present a data augmentation and retraining approach based on adversarial learning for improving anomaly detection. In particular, we first define a method for gener- ating adversarial examples for anomaly detectors based on Hidden Markov Models (HMMs). Then, we present a data augmentation and retraining technique that uses these adversarial examples to improve anomaly detection performance. Finally, we evaluate our adversarial data augmentation and retraining approach on four datasets showing that it achieves a statistically significant perfor- mance improvement and enhances the robustness to adversarial attacks. Key differences from the state-of-the-art on adversarial data augmentation are the focus on multivariate time series (as opposed to images), the context of one-class classification (in contrast to standard multi-class classification), and the use of HMMs (in contrast to neural networks)

    A Review of Adversarial Attacks in Computer Vision

    Full text link
    Deep neural networks have been widely used in various downstream tasks, especially those safety-critical scenario such as autonomous driving, but deep networks are often threatened by adversarial samples. Such adversarial attacks can be invisible to human eyes, but can lead to DNN misclassification, and often exhibits transferability between deep learning and machine learning models and real-world achievability. Adversarial attacks can be divided into white-box attacks, for which the attacker knows the parameters and gradient of the model, and black-box attacks, for the latter, the attacker can only obtain the input and output of the model. In terms of the attacker's purpose, it can be divided into targeted attacks and non-targeted attacks, which means that the attacker wants the model to misclassify the original sample into the specified class, which is more practical, while the non-targeted attack just needs to make the model misclassify the sample. The black box setting is a scenario we will encounter in practice

    On modeling and optimisation of air Traffic flow management problem with en-route capacities.

    Get PDF
    Master of Science in Mathematics, Statistics and Computer Science. University of KwaZulu-Natal, Durban 2016.The air transportation industry in the past ten years witnessed an upsurge with the number of passengers swelling exponentially. This development has seen a high demand in airport and airspace usage, which consequently has an enormous strain on the aviation industry of a given country. Although increase in airport capacity would be logical to meet this demand, factors such as poor weather conditions and other unforeseen ones have made it difficult if not impossible to do such. In fact there is a high probability of capacity reduction in most of the airports and air sectors within these regions. It is no surprise therefore that, most countries experience congestion almost on a daily basis. Congestion interrupts activities in the air transportation network and this has dire consequences on the air traffic control system as well as the nation's economy due to the significant costs incurred by airlines and passengers. This is against a background where most air tra c managers are met with the challenge of finding optimal scheduling strategies that can minimise delay costs. Current practices and research has shown that there is a high possibility of reducing the effects of congestion problems on the air traffic control system as well as the total delay costs incurred to the nearest minimum through an optimal control of ights. Optimal control of these ights can either be achieved by assigning ground holding delays or air borne delays together with any other control actions to mitigate congestion. This exposes a need for adequate air traffic ow management given that it plays a crucial role in alleviating delay costs. Air Traffic Flow Management (ATFM) is defined as a set of strategic processes that reduce air traffic delays and congestion problems. More precisely, it is the regulation of air traffic in such a way that the available airport and airspace capacity are utilised efficiently without been exceeded when handling traffic. The problem of managing air traffic so as to ensure efficient and safe ow of aircraft throughout the airspace is often referred to as the Air Traffic Flow Management Problem (ATFMP). This thesis provides a detailed insight on the ATFMP wherein the existing approaches, methodologies and optimisation techniques that have been (and continue to be) used to address the ATFMP were critically examined. Particular attention to optimisation models on airport capacity and airspace allocation were also discussed extensively as they depict what is obtainable in the air transportation system. Furthermore, the thesis attempted a comprehensive and, up-to-date review which extensively fed off literature on ATFMP. The instances in this literature were mainly derived from North America, Europe and Africa. Having reviewed the current ATFM practices and existing optimisation models and approaches for solving the ATFMP, the generalised basic model was extended to account for additional modeling variations. Furthermore, deterministic integer programming formulations were developed for reducing the air traffic delays and congestion problems based on the sector and path-based approaches already proposed for incorporating rerouting options into the basic ATFMP model. The formulation does not only takes into account all the ight phases but it also solves for optimal synthesis of other ow management activities including rerouting decisions, ight cancellation and penalisation. The claims from the basic ATFMP model was validated on artificially constructed datasets and generated instances. The computational performance of the basic and modified ATFMP reveals that the resulting solutions are completely integral, and an optimal solution can be obtained within the shortest possible computational time. Thereby, affirming the fact that these models can be used in effective decision making and efficient management of the air traffic flow

    What to Read: A Biased Guide to AI Literacy for the Beginner

    Get PDF
    Acknowledgements. It was Ken Forbus' idea, and he, Howie Shrobe, Dan Weld, and John Batali read various drafts. Dan Huttenlocher and Tom Knight helped with the speech recognition section. The science fiction section was prepared with the aid of my SF/AI editorial board, consisting of Carl Feynman and David Wallace, and of the ArpaNet SF-Lovers community. Even so, all responsibility rests with me.This note tries to provide a quick guide to AI literacy for the beginning AI hacker and for the experienced AI hacker or two whose scholarship isn't what it should be. most will recognize it as the same old list of classic papers, give or take a few that I feel to be under- or over-rated. It is not guaranteed to be thorough or balanced or anything like that.MIT Artificial Intelligence Laborator

    Weather2vec: Representation Learning for Causal Inference with Non-Local Confounding in Air Pollution and Climate Studies

    Full text link
    Estimating the causal effects of a spatially-varying intervention on a spatially-varying outcome may be subject to non-local confounding (NLC), a phenomenon that can bias estimates when the treatments and outcomes of a given unit are dictated in part by the covariates of other nearby units. In particular, NLC is a challenge for evaluating the effects of environmental policies and climate events on health-related outcomes such as air pollution exposure. This paper first formalizes NLC using the potential outcomes framework, providing a comparison with the related phenomenon of causal interference. Then, it proposes a broadly applicable framework, termed "weather2vec", that uses the theory of balancing scores to learn representations of non-local information into a scalar or vector defined for each observational unit, which is subsequently used to adjust for confounding in conjunction with causal inference methods. The framework is evaluated in a simulation study and two case studies on air pollution where the weather is an (inherently regional) known confounder
    • …
    corecore