687 research outputs found

    Deep learning for automobile predictive maintenance under Industry 4.0

    Get PDF
    Industry 4.0 refers to the fourth industrial revolution, which has boosted the development of the world. An important target of Industry 4.0 is to maximize the asset uptime so to improve productivity and reduce the production and maintenance cost. The emerging techniques such as artificial intelligence (AI), industrial Internet of things (IIoT) and cyber-physical system (CPS) have accelerated the development of data-orientated application such as predictive maintenance (PdM). Maintenance is a big concern for an automobile fleet management company. An accurate maintenance prediction can be helpful to avoid critical failure and avoid further loss. Deep learning is a type of prevailing machine learning algorithm which has been widely used in big data analytics. However, how to establish a maintenance prediction model based on historical maintenance data using deep learning has not been investigated. Moreover, it is worthwhile to study how to build a prediction model when the labelled data is insufficient. Furthermore, surrounding factors which may impact automobile lifecycle have not been concerned in the state-of-the-art. Hence, this thesis will focus on how to pave the way for automobile PdM under Industry 4.0. This research is structured according to four themes. Firstly, different from the conventional PdM research that only focuses on modelling based on sensor data or historical maintenance data, a framework for automobile PdM based on multi-source data is proposed. The proposed framework aims at automobile TBF modelling, prediction, and decision support based on the multi-source data. There are five layers designed in this framework, which are data collection, cloud data transmission and storage, data mapping, pre-processing and integration, deep learning for automobile TBF modelling, and decision support for PdM. This framework covers the entire knowledge discovery process from data collection to decision support. Secondly, one of the purposes of this thesis is to establish a Time-Between-Failure (TBF) prediction model through a data-driven approach. An accurate automobile TBF iv Abstract prediction can bring tangible benefits to a fleet management company. Different from the existing studies that adopted sensor data for failure time prediction, a new approach called Cox proportional hazard deep learning (CoxPHDL) is proposed based on the historical maintenance data for TBF modelling and prediction. CoxPHDL is able to tackle the data sparsity and data censoring issues that are common in the analysis of historical maintenance data. Firstly, an autoencoder is adopted to convert the nominal data into a robust representation. Secondly, a Cox PHM is researched to estimate the TBF of the censored data. A long-short-term memory (LSTM) network is then established to train the TBF prediction model based on the pre-processed maintenance data. Experimental results have demonstrated the merits of the proposed approach. Thirdly, a large amount of labelled data is one of the critical factors to the satisfactory algorithm performance of deep learning. However, labelled data is expensive to collect in the real world. In order to build a TBF prediction model using deep learning when the labelled data is limited, a new semi-supervised learning algorithm called deep learning embedded semi-supervised learning (DLeSSL) is proposed. Based on DLeSSL, unlabelled data can be estimated using a semi-supervised learning approach that has a deep learning technique embedded so to expand the labelled dataset. Results derived using the proposed method reveal that deep learning (DLeSSL based) outperforms the benchmarking algorithms when the labelled data is limited. In addition, different from existing studies, the effect on algorithm performance due to the size of labelled data and unlabelled data is reported to offer insights for the deployment of DLeSSL. Finally, automobile lifecycle can be impacted by surrounding factors such as weather, traffic, and terrain. The data contains these factors can be collected and processed via geographical information system (GIS). To introduce these GIS data into automobile TBF modelling, an integrated approach is proposed. This is the first time that the surrounding factors are considered in the study of automobile TBF modelling. Meanwhile, in order to build a TBF prediction model based on multi-source data, a new deep learning architecture called merged-LSTM (M-LSTM) network is designed. Abstract v Experimental results derived using the proposed approach and M-LSTM network reveal the impacts of the GIS factors. This thesis aims to research automobile PdM using deep learning, which provides a feasibility study for achieving Industry 4.0. As such, it offers great potential as a route to achieving a more profitable, efficient, and sustainable fleet management

    Unsupervised Automatic Detection Of Transient Phenomena In InSAR Time-Series using Machine Learning

    Get PDF
    The detection and measurement of transient episodes of crustal deformation from global InSAR datasets are crucial for a wide range of solid earth and natural hazard applications. But the large volumes of unlabelled data captured by satellites preclude manual systematic analysis, and the small signal-to-noise ratio makes the task difficult. In this thesis, I present a state-of-the-art, unsupervised and event-agnostic deep-learning based approach for the automatic identification of transient deformation events in noisy time-series of unwrapped InSAR images. I adopt an anomaly detection framework that learns the ‘normal’ spatio-temporal pattern of noise in the data, and which therefore identifies any transient deformation phenomena that deviate from this pattern as ‘anomalies’. The deep-learning model is built around a bespoke autoencoder that includes convolutional and LSTM layers, as well as a neural network which acts as a bridge between the encoder and decoder. I train our model on real InSAR data from northern Turkey and find it has an overall accuracy and true positive rate of around 85% when trying to detect synthetic deformation signals of length-scale > 350 m and magnitude > 4 cm. Furthermore, I also show the method can detect (1) a real Mw 5.7 earthquake in InSAR data from an entirely different region- SW Turkey, (2) a volcanic deformation in Domuyo, Argentina, (3) a synthetic slow-slip event and (4) an interseismic deformation around NAF in a descending frame in northern Turkey. Overall I show that my method is suitable for automated analysis of large, global InSAR datasets, and for robust detection and separation of deformation signals from nuisance signals in InSAR data

    AI-Assisted Cotton Grading: Active and Semi-Supervised Learning to Reduce the Image-Labelling Burden

    Get PDF
    The assessment of food and industrial crops during harvesting is important to determine the quality and downstream processing requirements, which in turn affect their market value. While machine learning models have been developed for this purpose, their deployment is hindered by the high cost of labelling the crop images to provide data for model training. This study examines the capabilities of semi-supervised and active learning to minimise effort when labelling cotton lint samples while maintaining high classification accuracy. Random forest classification models were developed using supervised learning, semi-supervised learning, and active learning to determine Egyptian cotton grade. Compared to supervised learning (80.20-82.66%) and semi-supervised learning (81.39-85.26%), active learning models were able to achieve higher accuracy (82.85-85.33%) with up to 46.4% reduction in the volume of labelled data required. The primary obstacle when using machine learning for Egyptian cotton grading is the time required for labelling cotton lint samples. However, by applying active learning, this study successfully decreased the time needed from 422.5 to 177.5 min. The findings of this study demonstrate that active learning is a promising approach for developing accurate and efficient machine learning models for grading food and industrial crops

    When Less is More: On the Value of "Co-training" for Semi-Supervised Software Defect Predictors

    Full text link
    Labeling a module defective or non-defective is an expensive task. Hence, there are often limits on how much-labeled data is available for training. Semi-supervised classifiers use far fewer labels for training models. However, there are numerous semi-supervised methods, including self-labeling, co-training, maximal-margin, and graph-based methods, to name a few. Only a handful of these methods have been tested in SE for (e.g.) predicting defects and even there, those methods have been tested on just a handful of projects. This paper applies a wide range of 55 semi-supervised learners to over 714 projects. We find that semi-supervised "co-training methods" work significantly better than other approaches. Specifically, after labeling, just 2.5% of data, then make predictions that are competitive to those using 100% of the data. That said, co-training needs to be used cautiously since the specific choice of co-training methods needs to be carefully selected based on a user's specific goals. Also, we warn that a commonly-used co-training method ("multi-view"-- where different learners get different sets of columns) does not improve predictions (while adding too much to the run time costs 11 hours vs. 1.8 hours). It is an open question, worthy of future work, to test if these reductions can be seen in other areas of software analytics. To assist with exploring other areas, all the codes used are available at https://github.com/ai-se/Semi-Supervised.Comment: 36 pages, 10 figures, 5 table

    Machine learning with limited label availability: algorithms and applications

    Get PDF
    L'abstract è presente nell'allegato / the abstract is in the attachmen
    • …
    corecore