1,429 research outputs found

    Fault detection in operating helicopter drive train components based on support vector data description

    Get PDF
    The objective of the paper is to develop a vibration-based automated procedure dealing with early detection of mechanical degradation of helicopter drive train components using Health and Usage Monitoring Systems (HUMS) data. An anomaly-detection method devoted to the quantification of the degree of deviation of the mechanical state of a component from its nominal condition is developed. This method is based on an Anomaly Score (AS) formed by a combination of a set of statistical features correlated with specific damages, also known as Condition Indicators (CI), thus the operational variability is implicitly included in the model through the CI correlation. The problem of fault detection is then recast as a one-class classification problem in the space spanned by a set of CI, with the aim of a global differentiation between normal and anomalous observations, respectively related to healthy and supposedly faulty components. In this paper, a procedure based on an efficient one-class classification method that does not require any assumption on the data distribution, is used. The core of such an approach is the Support Vector Data Description (SVDD), that allows an efficient data description without the need of a significant amount of statistical data. Several analyses have been carried out in order to validate the proposed procedure, using flight vibration data collected from a H135, formerly known as EC135, servicing helicopter, for which micro-pitting damage on a gear was detected by HUMS and assessed through visual inspection. The capability of the proposed approach of providing better trade-off between false alarm rates and missed detection rates with respect to individual CI and to the AS obtained assuming jointly-Gaussian-distributed CI has been also analysed

    Thirty Years of Machine Learning: The Road to Pareto-Optimal Wireless Networks

    Full text link
    Future wireless networks have a substantial potential in terms of supporting a broad range of complex compelling applications both in military and civilian fields, where the users are able to enjoy high-rate, low-latency, low-cost and reliable information services. Achieving this ambitious goal requires new radio techniques for adaptive learning and intelligent decision making because of the complex heterogeneous nature of the network structures and wireless services. Machine learning (ML) algorithms have great success in supporting big data analytics, efficient parameter estimation and interactive decision making. Hence, in this article, we review the thirty-year history of ML by elaborating on supervised learning, unsupervised learning, reinforcement learning and deep learning. Furthermore, we investigate their employment in the compelling applications of wireless networks, including heterogeneous networks (HetNets), cognitive radios (CR), Internet of things (IoT), machine to machine networks (M2M), and so on. This article aims for assisting the readers in clarifying the motivation and methodology of the various ML algorithms, so as to invoke them for hitherto unexplored services as well as scenarios of future wireless networks.Comment: 46 pages, 22 fig

    Featured Anomaly Detection Methods and Applications

    Get PDF
    Anomaly detection is a fundamental research topic that has been widely investigated. From critical industrial systems, e.g., network intrusion detection systems, to people’s daily activities, e.g., mobile fraud detection, anomaly detection has become the very first vital resort to protect and secure public and personal properties. Although anomaly detection methods have been under consistent development over the years, the explosive growth of data volume and the continued dramatic variation of data patterns pose great challenges on the anomaly detection systems and are fuelling the great demand of introducing more intelligent anomaly detection methods with distinct characteristics to cope with various needs. To this end, this thesis starts with presenting a thorough review of existing anomaly detection strategies and methods. The advantageous and disadvantageous of the strategies and methods are elaborated. Afterward, four distinctive anomaly detection methods, especially for time series, are proposed in this work aiming at resolving specific needs of anomaly detection under different scenarios, e.g., enhanced accuracy, interpretable results, and self-evolving models. Experiments are presented and analysed to offer a better understanding of the performance of the methods and their distinct features. To be more specific, the abstracts of the key contents in this thesis are listed as follows: 1) Support Vector Data Description (SVDD) is investigated as a primary method to fulfill accurate anomaly detection. The applicability of SVDD over noisy time series datasets is carefully examined and it is demonstrated that relaxing the decision boundary of SVDD always results in better accuracy in network time series anomaly detection. Theoretical analysis of the parameter utilised in the model is also presented to ensure the validity of the relaxation of the decision boundary. 2) To support a clear explanation of the detected time series anomalies, i.e., anomaly interpretation, the periodic pattern of time series data is considered as the contextual information to be integrated into SVDD for anomaly detection. The formulation of SVDD with contextual information maintains multiple discriminants which help in distinguishing the root causes of the anomalies. 3) In an attempt to further analyse a dataset for anomaly detection and interpretation, Convex Hull Data Description (CHDD) is developed for realising one-class classification together with data clustering. CHDD approximates the convex hull of a given dataset with the extreme points which constitute a dictionary of data representatives. According to the dictionary, CHDD is capable of representing and clustering all the normal data instances so that anomaly detection is realised with certain interpretation. 4) Besides better anomaly detection accuracy and interpretability, better solutions for anomaly detection over streaming data with evolving patterns are also researched. Under the framework of Reinforcement Learning (RL), a time series anomaly detector that is consistently trained to cope with the evolving patterns is designed. Due to the fact that the anomaly detector is trained with labeled time series, it avoids the cumbersome work of threshold setting and the uncertain definitions of anomalies in time series anomaly detection tasks

    Unsupervised Similarity-Based Risk Stratification for Cardiovascular Events Using Long-Term Time-Series Data

    Get PDF
    In medicine, one often bases decisions upon a comparative analysis of patient data. In this paper, we build upon this observation and describe similarity-based algorithms to risk stratify patients for major adverse cardiac events. We evolve the traditional approach of comparing patient data in two ways. First, we propose similarity-based algorithms that compare patients in terms of their long-term physiological monitoring data. Symbolic mismatch identifies functional units in long-term signals and measures changes in the morphology and frequency of these units across patients. Second, we describe similarity-based algorithms that are unsupervised and do not require comparisons to patients with known outcomes for risk stratification. This is achieved by using an anomaly detection framework to identify patients who are unlike other patients in a population and may potentially be at an elevated risk. We demonstrate the potential utility of our approach by showing how symbolic mismatch-based algorithms can be used to classify patients as being at high or low risk of major adverse cardiac events by comparing their long-term electrocardiograms to that of a large population. We describe how symbolic mismatch can be used in three different existing methods: one-class support vector machines, nearest neighbor analysis, and hierarchical clustering. When evaluated on a population of 686 patients with available long-term electrocardiographic data, symbolic mismatch-based comparative approaches were able to identify patients at roughly a two-fold increased risk of major adverse cardiac events in the 90 days following acute coronary syndrome. These results were consistent even after adjusting for other clinical risk variables.National Science Foundation (U.S.) (CAREER award 1054419

    Flexible Time Series Matching for Clinical and Behavioral Data

    Get PDF
    Time Series data became broadly applied by the research community in the last decades after a massive explosion of its availability. Nonetheless, this rise required an improvement in the existing analysis techniques which, in the medical domain, would help specialists to evaluate their patients condition. One of the key tasks in time series analysis is pattern recognition (segmentation and classification). Traditional methods typically perform subsequence matching, making use of a pattern template and a similarity metric to search for similar sequences throughout time series. However, real-world data is noisy and variable (morphological distortions), making a template-based exact matching an elementary approach. Intending to increase flexibility and generalize the pattern searching tasks across domains, this dissertation proposes two Deep Learning-based frameworks to solve pattern segmentation and anomaly detection problems. Regarding pattern segmentation, a Convolution/Deconvolution Neural Network is proposed, learning to distinguish, point-by-point, desired sub-patterns from background content within a time series. The proposed framework was validated in two use-cases: electrocardiogram (ECG) and inertial sensor-based human activity (IMU) signals. It outperformed two conventional matching techniques, being capable of notably detecting the targeted cycles even in noise-corrupted or extremely distorted signals, without using any reference template nor hand-coded similarity scores. Concerning anomaly detection, the proposed unsupervised framework uses the reconstruction ability of Variational Autoencoders and a local similarity score to identify non-labeled abnormalities. The proposal was validated in two public ECG datasets (MITBIH Arrhythmia and ECG5000), performing cardiac arrhythmia identification. Results indicated competitiveness relative to recent techniques, achieving detection AUC scores of 98.84% (ECG5000) and 93.32% (MIT-BIH Arrhythmia).Dados de séries temporais tornaram-se largamente aplicados pela comunidade científica nas últimas decadas após um aumento massivo da sua disponibilidade. Contudo, este aumento exigiu uma melhoria das atuais técnicas de análise que, no domínio clínico, auxiliaria os especialistas na avaliação da condição dos seus pacientes. Um dos principais tipos de análise em séries temporais é o reconhecimento de padrões (segmentação e classificação). Métodos tradicionais assentam, tipicamente, em técnicas de correspondência em subsequências, fazendo uso de um padrão de referência e uma métrica de similaridade para procurar por subsequências similares ao longo de séries temporais. Todavia, dados do mundo real são ruidosos e variáveis (morfologicamente), tornando uma correspondência exata baseada num padrão de referência uma abordagem rudimentar. Pretendendo aumentar a flexibilidade da análise de séries temporais e generalizar tarefas de procura de padrões entre domínios, esta dissertação propõe duas abordagens baseadas em Deep Learning para solucionar problemas de segmentação de padrões e deteção de anomalias. Acerca da segmentação de padrões, a rede neuronal de Convolução/Deconvolução proposta aprende a distinguir, ponto a ponto, sub-padrões pretendidos de conteúdo de fundo numa série temporal. O modelo proposto foi validado em dois casos de uso: sinais eletrocardiográficos (ECG) e de sensores inerciais em atividade humana (IMU). Este superou duas técnicas convencionais, sendo capaz de detetar os ciclos-alvo notavelmente, mesmo em sinais corrompidos por ruído ou extremamente distorcidos, sem o uso de nenhum padrão de referência nem métricas de similaridade codificadas manualmente. A respeito da deteção de anomalias, a técnica não supervisionada proposta usa a capacidade de reconstrução dos Variational Autoencoders e uma métrica de similaridade local para identificar anomalias desconhecidas. A proposta foi validada na identificação de arritmias cardíacas em duas bases de dados públicas de ECG (MIT-BIH Arrhythmia e ECG5000). Os resultados revelam competitividade face a técnicas recentes, alcançando métricas AUC de deteção de 93.32% (MIT-BIH Arrhythmia) e 98.84% (ECG5000)

    Transforming Data into Meaning. Data-Driven approaches for Particle Physics, Nuclear Power Safety and Humanitarian Crisis Situations.

    Get PDF
    Machine learning and data intensive methods can be applied to a plethora of research domains. We apply supervised and unsupervised machine learning, Monte Carlo simulations and statistical tools to three diverse areas of research, tackling a range of computational and data analysis challenges unique to their respective environments. Using SHERPA-a Monte Carlo event generator-as a Standard Model machine we generate thousands of particle collision events. We employ a range of neural network architectures to determine the most powerful discriminating features which eliminate vast numbers of background events enabling us to calculate new constraints on the charm Yukawa coupling at the Large Hadron Collider and future projections. Hartlepool Nuclear Power Station has a rich array of instrumentation that continuously monitors reactor health as frequently as every second, at all times. We apply unsupervised machine learning and Bayesian tools to scrutinise anomalous behaviour in the data which is indicative of instrumentation degradation prior to instrumentation failure. JUNE-an agent based epidemiological simulation-is used to extract novel social mixing matrices at Cox's Bazar, a refugee camp in Bangladesh containing ∼600,000{\sim}600,000 displaced people. These contact matrices can be used to understand social interactions and disease spread and therefore provide better utilisation of limited resources

    Information Theory and Its Application in Machine Condition Monitoring

    Get PDF
    Condition monitoring of machinery is one of the most important aspects of many modern industries. With the rapid advancement of science and technology, machines are becoming increasingly complex. Moreover, an exponential increase of demand is leading an increasing requirement of machine output. As a result, in most modern industries, machines have to work for 24 hours a day. All these factors are leading to the deterioration of machine health in a higher rate than before. Breakdown of the key components of a machine such as bearing, gearbox or rollers can cause a catastrophic effect both in terms of financial and human costs. In this perspective, it is important not only to detect the fault at its earliest point of inception but necessary to design the overall monitoring process, such as fault classification, fault severity assessment and remaining useful life (RUL) prediction for better planning of the maintenance schedule. Information theory is one of the pioneer contributions of modern science that has evolved into various forms and algorithms over time. Due to its ability to address the non-linearity and non-stationarity of machine health deterioration, it has become a popular choice among researchers. Information theory is an effective technique for extracting features of machines under different health conditions. In this context, this book discusses the potential applications, research results and latest developments of information theory-based condition monitoring of machineries
    • …
    corecore