582 research outputs found

    Automated detection of vehicles with machine learning

    Get PDF
    Considering the significant volume of data generated by sensor systems and network hardware which is required to be analysed and interpreted by security analysts, the potential for human error is significant. This error can lead to consequent harm for some systems in the event of an adverse event not being detected. In this paper we compare two machine learning algorithms that can assist in supporting the security function effectively and present results that can be used to select the best algorithm for a specific domain. It is suggested that a naïve Bayesian classifier (NBC) and an artificial neural network (ANN) are most likely the best candidate algorithms for the proposed application. It was found that the NBC was faster and more accurate than the ANN for the given data set. Future research will look to repeat this process for cyber security specific applications, and also examine GPGPU optimisations to the machine learning algorithms.

    Automated Detection of Vehicles with Machine Learning

    Get PDF
    Considering the significant volume of data generated by sensor systems and network hardware which is required to be analysed and intepreted by security analysts, the potential for human error is significant. This error can lead to consequent harm for some systems in the event of an adverse event not being detected. In this paper we compare two machine learning algorithms that can assist in supporting the security function effectively and present results that can be used to select the best algorithm for a specific domain. It is suggested that a naive Bayesian classiifer (NBC) and an artificial neural network (ANN) are most likely the best candidate algorithms for the proposed application. It was found that NBC was faster and more accurate than the ANN for the given data set. Future research will look to repeat this process for cyber security specific applications, and also examine teh GPGPU optimisations to the machine learning algorithms

    Euclidean Space Data Projection Classifier with Cartesian Genetic Programming (CGP)

    Get PDF
    Most evolutionary based classifiers are built based on generated rules sets that categorize the data into respective classes. This research work is a preliminary work which proposes an evolutionary-based classifier using a simplified Cartesian Genetic Programming (CGP) evolutionary algorithm. Instead on using evolutionary generated rule sets, the CGP generates i) a reference coordinate ii) projection functions to project data into a new 3 Dimensional Euclidean space. Subsequently, a distance boundary function of the new projected data to the reference coordinates is applied to classify the data into their respective classes. The evolutionary algorithm is based on a simplified CGP Algorithm using a 1+4 evolutionary strategy. The data projection functions were evolved using CGP for 1000 generations before stopping to extract the best functions. The Classifier was tested using three PROBEN 1 benchmarking datasets which are the PIMA Indians diabetes dataset, Heart Disease dataset and Wisconsin Breast Cancer (WBC) Dataset based on 10 fold cross validation dataset partitioning. Testing results showed that data projection function generated competitive results classification rates: Cancer dataset (97.71%), PIMA Indians dataset (77.92%) and heart disease (85.86%)

    Automated Analysis of Mammograms using Evolutionary Algorithms

    Get PDF
    Breast cancer is the leading cause of death in women in the western countries. The diagnosis of breast cancer at the earlier stage may be particularly important since it provides early treatment, this will decreases the chance of cancer spreading and increase the survival rates. The hard work is the early detection of any tissues abnormal and confirmation of their cancerous natures. In additionally, finding abnormal on very early stage can also affected by poor quality of image and other problems that might show on a mammogram. Mammograms are high resolution x-rays of the breast that are widely used to screen for cancer in women. This report describes the stages of development of a novel representation of Cartesian Genetic programming as part of a computer aided diagnosis system. Specifically, this work is concerned with automated recognition of microcalcifications, one of the key structures used to identify cancer. Results are presented for the application of the proposed algorithm to a number of mammogram sections taken from the Lawrence Livermore National Laboratory Database. The performance of any algorithm such as evolutionary algorithm is only good as the data it is trained on. More specifically, the class represented in the training data must consist of the true examples or else reliable classifications. Considering the difficulties in obtaining a previously constructed database, there is a new database has been construct to avoiding pitfalls and lead on the novel evolutional algorithm Multi-chromosome Cartesian genetic programming the success on classification of microcalcifications in mammograms

    Layered genetic programming for feature extraction in classification problems

    Get PDF
    Dissertation presented as the partial requirement for obtaining a Master's degree in Data Science and Advanced AnalyticsGenetic programming has been proven to be a successful technique for feature extraction in various applications. In this thesis, we present a Layered Genetic Programming system which implements genetic programming-based feature extraction mechanism. The proposed system uses a layered structure where instead of evolving just one population of individuals, several populations are evolved sequentially. Each such population transforms the input data received from the previous population into a lower dimensional space with the aim of improving classification performance. The performance of the proposed system was experimentally tested on 5 real-world problems using different dimensionality reduction step sizes and different classifiers. The proposed method was able to outperform a simple classifier applied directly on the original data on two problems. On the remaining problems, the classifier performed better using the original data. The best solutions were often obtained in the first few layers which implied that increasing the size of the system, i.e. adding more layers was not useful. However, the layered structure allowed control of the size of individuals

    Evolving Artificial Neural Networks using Cartesian Genetic Programming

    Get PDF
    NeuroEvolution is the application of Evolutionary Algorithms to the training of Artificial Neural Networks. NeuroEvolution is thought to possess many benefits over traditional training methods including: the ability to train recurrent network structures, the capability to adapt network topology, being able to create heterogeneous networks of arbitrary transfer functions, and allowing application to reinforcement as well as supervised learning tasks. This thesis presents a series of rigorous empirical investigations into many of these perceived advantages of NeuroEvolution. In this work it is demonstrated that the ability to simultaneously adapt network topology along with connection weights represents a significant advantage of many NeuroEvolutionary methods. It is also demonstrated that the ability to create heterogeneous networks comprising a range of transfer functions represents a further significant advantage. This thesis also investigates many potential benefits and drawbacks of NeuroEvolution which have been largely overlooked in the literature. This includes the presence and role of genetic redundancy in NeuroEvolution's search and whether program bloat is a limitation. The investigations presented focus on the use of a recently developed NeuroEvolution method based on Cartesian Genetic Programming. This thesis extends Cartesian Genetic Programming such that it can represent recurrent program structures allowing for the creation of recurrent Artificial Neural Networks. Using this newly developed extension, Recurrent Cartesian Genetic Programming, and its application to Artificial Neural Networks, are demonstrated to be extremely competitive in the domain of series forecasting

    Kartezio: Evolutionary Design of Explainable Pipelines for Biomedical Image Analysis

    Full text link
    An unresolved issue in contemporary biomedicine is the overwhelming number and diversity of complex images that require annotation, analysis and interpretation. Recent advances in Deep Learning have revolutionized the field of computer vision, creating algorithms that compete with human experts in image segmentation tasks. Crucially however, these frameworks require large human-annotated datasets for training and the resulting models are difficult to interpret. In this study, we introduce Kartezio, a modular Cartesian Genetic Programming based computational strategy that generates transparent and easily interpretable image processing pipelines by iteratively assembling and parameterizing computer vision functions. The pipelines thus generated exhibit comparable precision to state-of-the-art Deep Learning approaches on instance segmentation tasks, while requiring drastically smaller training datasets, a feature which confers tremendous flexibility, speed, and functionality to this approach. We also deployed Kartezio to solve semantic and instance segmentation problems in four real-world Use Cases, and showcase its utility in imaging contexts ranging from high-resolution microscopy to clinical pathology. By successfully implementing Kartezio on a portfolio of images ranging from subcellular structures to tumoral tissue, we demonstrated the flexibility, robustness and practical utility of this fully explicable evolutionary designer for semantic and instance segmentation.Comment: 36 pages, 6 main Figures. The Extended Data Movie is available at the following link: https://www.youtube.com/watch?v=r74gdzb6hdA. The source code is available on Github: https://github.com/KevinCortacero/Kartezi

    Advancement in Research Techniques on Medical Imaging Processing for Breast Cancer Detection

    Get PDF
    With the advancement of medical image processing, the area of the healthcare sector has started receiving the benefits of the modern arena of diagnostic tools to identify the diseases effectively. Cancer is one of the dreaded diseases, where success factor of treatment offered by medical sector is still an unsolved problem. Hence, the success factor of the treatment lies in early stage of the disease or timely detection of the disease. This paper discusses about the advancement being made in the medical image processing towards an effective diagnosis of the breast cancer from the mammogram image in radiology. There has been enough research activity with various sorts of advances techniques being implemented in the past decade. The prime contribution of this manuscript is to showcase the advancement of the technology along with illustration of the effectiveness of the existing literatures with respect to research gap

    Automated Feature Engineering for Deep Neural Networks with Genetic Programming

    Get PDF
    Feature engineering is a process that augments the feature vector of a machine learning model with calculated values that are designed to enhance the accuracy of a model’s predictions. Research has shown that the accuracy of models such as deep neural networks, support vector machines, and tree/forest-based algorithms sometimes benefit from feature engineering. Expressions that combine one or more of the original features usually create these engineered features. The choice of the exact structure of an engineered feature is dependent on the type of machine learning model in use. Previous research demonstrated that various model families benefit from different types of engineered feature. Random forests, gradient-boosting machines, or other tree-based models might not see the same accuracy gain that an engineered feature allowed neural networks, generalized linear models, or other dot-product based models to achieve on the same data set. This dissertation presents a genetic programming-based algorithm that automatically engineers features that increase the accuracy of deep neural networks for some data sets. For a genetic programming algorithm to be effective, it must prioritize the search space and efficiently evaluate what it finds. This dissertation algorithm faced a potential search space composed of all possible mathematical combinations of the original feature vector. Five experiments were designed to guide the search process to efficiently evolve good engineered features. The result of this dissertation is an automated feature engineering (AFE) algorithm that is computationally efficient, even though a neural network is used to evaluate each candidate feature. This approach gave the algorithm a greater opportunity to specifically target deep neural networks in its search for engineered features that improve accuracy. Finally, a sixth experiment empirically demonstrated the degree to which this algorithm improved the accuracy of neural networks on data sets augmented by the algorithm’s engineered features
    • …
    corecore