457 research outputs found

    A review on the self and dual interactions between machine learning and optimisation

    Get PDF
    Machine learning and optimisation are two growing fields of artificial intelligence with an enormous number of computer science applications. The techniques in the former area aim to learn knowledge from data or experience, while the techniques from the latter search for the best option or solution to a given problem. To employ these techniques automatically and effectively aligning with the real aim of artificial intelligence, both sets of techniques are frequently hybridised, interacting with each other and themselves. This study focuses on such interactions aiming at (1) presenting a broad overview of the studies on self and dual interactions between machine learning and optimisation; (2) providing a useful tutorial for researchers and practitioners in both fields in support of collaborative work through investigation of the recent advances and analyses of the advantages and disadvantages of different techniques to tackle the same or similar problems; (3) clarifying the overlapping terminologies having different meanings used in both fields; (4) identifying research gaps and potential research directions

    Decision making with reciprocal chains and binary neural network models

    Get PDF
    Automated decision making systems are relied on in increasingly diverse and critical settings. Human users expect such systems to improve or augment their own decision making in complex scenarios, in real time, often across distributed networks of devices. This thesis studies binary decision making systems of two forms. The rst system is built from a reciprocal chain, a statistical model able to capture the intentional behaviour of targets moving through a statespace, such as moving towards a destination state. The rst part of the thesis questions the utility of this higher level information in a tracking problem where the system must decide whether a target exists or not. The contributions of this study characterise the bene ts to be expected from reciprocal chains for tracking, using statistical tools and a novel simulation environment that provides relevant numerical experiments. Real world decision making systems often combine statistical models, such as the reciprocal chain, with the second type of system studied in this thesis, a neural network. In the tracking context, a neural network typically forms the object detection system. However, the power consumption and memory usage of state of the art neural networks makes their use on small devices infeasible. This motivates the study of binary neural networks in the second part of the thesis. Such networks use less memory and are e cient to run, compared to standard full precision networks. However, their optimisation is di cult, due to the non-di erentiable functions involved. Several algorithms elect to optimise surrogate networks that are di erentiable and correspond in some way to the original binary network. Unfortunately, the many choices involved in the algorithm design are poorly understood. The second part of the thesis questions the role of parameter initialisation in the optimisation of binary neural networks. Borrowing analytic tools from statistical physics, it is possible to characterise the typical behaviour of a range of algorithms at initialisation precisely, by studying how input signals propagate through these networks on average. This theoretical development also yields practical outcomes, providing scales that limit network depth and suggesting new initialisation methods for binary neural networks.Thesis (Ph.D.) -- University of Adelaide, School of Electrical & Electronic Engineering, 202

    From metaheuristics to learnheuristics: Applications to logistics, finance, and computing

    Get PDF
    Un gran nombre de processos de presa de decisions en sectors estratègics com el transport i la producció representen problemes NP-difícils. Sovint, aquests processos es caracteritzen per alts nivells d'incertesa i dinamisme. Les metaheurístiques són mètodes populars per a resoldre problemes d'optimització difícils en temps de càlcul raonables. No obstant això, sovint assumeixen que els inputs, les funcions objectiu, i les restriccions són deterministes i conegudes. Aquests constitueixen supòsits forts que obliguen a treballar amb problemes simplificats. Com a conseqüència, les solucions poden conduir a resultats pobres. Les simheurístiques integren la simulació a les metaheurístiques per resoldre problemes estocàstics d'una manera natural. Anàlogament, les learnheurístiques combinen l'estadística amb les metaheurístiques per fer front a problemes en entorns dinàmics, en què els inputs poden dependre de l'estructura de la solució. En aquest context, les principals contribucions d'aquesta tesi són: el disseny de les learnheurístiques, una classificació dels treballs que combinen l'estadística / l'aprenentatge automàtic i les metaheurístiques, i diverses aplicacions en transport, producció, finances i computació.Un gran número de procesos de toma de decisiones en sectores estratégicos como el transporte y la producción representan problemas NP-difíciles. Frecuentemente, estos problemas se caracterizan por altos niveles de incertidumbre y dinamismo. Las metaheurísticas son métodos populares para resolver problemas difíciles de optimización de manera rápida. Sin embargo, suelen asumir que los inputs, las funciones objetivo y las restricciones son deterministas y se conocen de antemano. Estas fuertes suposiciones conducen a trabajar con problemas simplificados. Como consecuencia, las soluciones obtenidas pueden tener un pobre rendimiento. Las simheurísticas integran simulación en metaheurísticas para resolver problemas estocásticos de una manera natural. De manera similar, las learnheurísticas combinan aprendizaje estadístico y metaheurísticas para abordar problemas en entornos dinámicos, donde los inputs pueden depender de la estructura de la solución. En este contexto, las principales aportaciones de esta tesis son: el diseño de las learnheurísticas, una clasificación de trabajos que combinan estadística / aprendizaje automático y metaheurísticas, y varias aplicaciones en transporte, producción, finanzas y computación.A large number of decision-making processes in strategic sectors such as transport and production involve NP-hard problems, which are frequently characterized by high levels of uncertainty and dynamism. Metaheuristics have become the predominant method for solving challenging optimization problems in reasonable computing times. However, they frequently assume that inputs, objective functions and constraints are deterministic and known in advance. These strong assumptions lead to work on oversimplified problems, and the solutions may demonstrate poor performance when implemented. Simheuristics, in turn, integrate simulation into metaheuristics as a way to naturally solve stochastic problems, and, in a similar fashion, learnheuristics combine statistical learning and metaheuristics to tackle problems in dynamic environments, where inputs may depend on the structure of the solution. The main contributions of this thesis include (i) a design for learnheuristics; (ii) a classification of works that hybridize statistical and machine learning and metaheuristics; and (iii) several applications for the fields of transport, production, finance and computing

    Automated design of population-based algorithms: a case study in vehicle routing

    Get PDF
    Metaheuristics have been extensively studied to solve constraint combinatorial optimisation problems such as vehicle routing problems. Most existing algorithms require considerable human effort and different kinds of expertise in algorithm design. These manually designed algorithms are discarded after solving the specific instances. It is highly desirable to automate the design of search algorithms, thus to solve problem instances effectively with less human intervention. This thesis develops a novel general search framework to formulate in a unified way a range of population-based algorithms. Within this framework, generic algorithmic components such as selection heuristics on the population and evolution operators are defined, and can be composed using machine learning to generate effective search algorithms automatically. This unified framework aims to serve as the basis to analyse algorithmic components, generating effective search algorithms for complex combinatorial optimisation problems. Three key research issues within the general search framework are identified: automated design of evolution operators, of selection heuristics, and of both. To accurately describe the search space of algorithm design as a new task for machine learning, this thesis identifies new key features, namely search-dependent and instance-dependent features. These features are identified to assist effective algorithm design. With these features, a set of state-of-the-art reinforcement learning techniques, such as deep Q-network based and proximal policy optimisation based models and maximum entropy mechanisms have been developed to intelligently select and combine appropriate evolution operators and selection heuristics during different stages of the optimisation process. The effectiveness and generality of these algorithms automatically designed within the proposed general search framework are validated comprehensively across different capacitated vehicle routing problem with time windows benchmark instances. This thesis contributes to making a key step towards automated algorithm design with a general framework supporting fundamental analysis by effective machine learning

    Automated design of population-based algorithms: a case study in vehicle routing

    Get PDF
    Metaheuristics have been extensively studied to solve constraint combinatorial optimisation problems such as vehicle routing problems. Most existing algorithms require considerable human effort and different kinds of expertise in algorithm design. These manually designed algorithms are discarded after solving the specific instances. It is highly desirable to automate the design of search algorithms, thus to solve problem instances effectively with less human intervention. This thesis develops a novel general search framework to formulate in a unified way a range of population-based algorithms. Within this framework, generic algorithmic components such as selection heuristics on the population and evolution operators are defined, and can be composed using machine learning to generate effective search algorithms automatically. This unified framework aims to serve as the basis to analyse algorithmic components, generating effective search algorithms for complex combinatorial optimisation problems. Three key research issues within the general search framework are identified: automated design of evolution operators, of selection heuristics, and of both. To accurately describe the search space of algorithm design as a new task for machine learning, this thesis identifies new key features, namely search-dependent and instance-dependent features. These features are identified to assist effective algorithm design. With these features, a set of state-of-the-art reinforcement learning techniques, such as deep Q-network based and proximal policy optimisation based models and maximum entropy mechanisms have been developed to intelligently select and combine appropriate evolution operators and selection heuristics during different stages of the optimisation process. The effectiveness and generality of these algorithms automatically designed within the proposed general search framework are validated comprehensively across different capacitated vehicle routing problem with time windows benchmark instances. This thesis contributes to making a key step towards automated algorithm design with a general framework supporting fundamental analysis by effective machine learning

    Decision support continuum paradigm for cardiovascular disease: Towards personalized predictive models

    Get PDF
    Clinical decision making is a ubiquitous and frequent task physicians make in their daily clinical practice. Conventionally, physicians adopt a cognitive predictive modelling process (i.e. knowledge and experience learnt from past lecture, research, literature, patients, etc.) for anticipating or ascertaining clinical problems based on clinical risk factors that they deemed to be most salient. However, with the inundation of health data and the confounding characteristics of diseases, more effective clinical prediction approaches are required to address these challenges. Approximately a few century ago, the first major transformation of medical practice took place as science-based approaches emerged with compelling results. Now, in the 21st century, new advances in science will once again transform healthcare. Data science has been postulated as an important component in this healthcare reform and has received escalating interests for its potential for ‘personalizing’ medicine. The key advantages of having personalized medicine include, but not limited to, (1) more effective methods for disease prevention, management and treatment, (2) improved accuracy for clinical diagnosis and prognosis, (3) provide patient-oriented personal health plan, and (4) cost containment. In view of the paramount importance of personalized predictive models, this thesis proposes 2 novel learning algorithms (i.e. an immune-inspired algorithm called the Evolutionary Data-Conscious Artificial Immune Recognition System, and a neural-inspired algorithm called the Artificial Neural Cell System for classification) and 3 continuum-based paradigms (i.e. biological, time and age continuum) for enhancing clinical prediction. Cardiovascular disease has been selected as the disease under investigation as it is an epidemic and major health concern in today’s world. We believe that our work has a meaningful and significant impact to the development of future healthcare system and we look forward to the wide adoption of advanced medical technologies by all care centres in the near future.Open Acces

    Applications of Hyper-parameter Optimisations for Static Malware Detection

    Get PDF
    Malware detection is a major security concern and a great deal of academic and commercial research and development is directed at it. Machine Learning is a natural technology to harness for malware detection and many researchers have investigated its use. However, drawing comparisons between different techniques is a fraught affair. For example, the performance of ML algorithms often depends significantly on parametric choices, so the question arises as to what parameter choices are optimal. In this thesis, we investigate the use of a variety of ML algorithms for building malware classifiers and also how best to tune the parameters of those algorithms – a process generally known as hyper-parameter optimisation (HPO). Firstly, we examine the effects of some simple (model-free) ways of parameter tuning together with a state-of-the-art Bayesian model-building approach. We demonstrate that optimal parameter choices may differ significantly from default choices and argue that hyper-parameter optimisation should be adopted as a ‘formal outer loop’ in the research and development of malware detection systems. Secondly, we investigate the use of covering arrays (combinatorial testing) as a way to combat the curse of dimensionality in Gird Search. Four ML techniques were used: Random Forests, xgboost, Light GBM and Decision Trees. cAgen (a tool that is used for combinatorial testing) is shown to be capable of generating high-performing subsets of the full parameter grid of Grid Search and so provides a rigorous but highly efficient means of performing HPO. This may be regarded as a ‘design of experiments’ approach. Thirdly, Evolutionary algorithms (EAs) were used to enhance machine learning classifier accuracy. Six traditional machine learning techniques baseline accuracy is recorded. Two evolutionary algorithm frameworks Tree-Based Pipeline Optimization Tool (TPOT) and Distributed Evolutionary Algorithms in Python (Deap) are compared. Deap shows very promising results for our malware detection problem. Fourthly, we compare the use of Grid Search and covering arrays for tuning the hyper-parameters of Neural Networks. Several major hyper-parameters were studied with various values and results. We achieve significant improvements over the benchmark model. Our work is carried out using EMBER, a major published malware benchmark dataset of Windows Portable Execution (PE) metadata samples, and a smaller dataset from kaggle.com (also comprising of Windows Portable Execution metadata). Overall, we conclude that HPO is an essential part of credible evaluations of ML-based malware detection models. We also demonstrate that high-performing hyper-parameter values can be found by HPO and that these can be found efficiently

    Adaptive and Scalable Controller Placement in Software-Defined Networking

    Get PDF
    Software-defined networking (SDN) revolutionizes network control by externalizing and centralizing the control plane. A critical aspect of SDN is Controller Placement (CP), which involves identifying the ideal number and location of controllers in a network to fulfill diverse objectives such as latency constraints (node-to-controller and controller-controller delay), fault tolerance, and controller load. Existing optimization techniques like Multi-Objective Particle Swarm Optimisation (MOPSO), Adapted Non-Dominating Sorting Genetic Algorithm-III (ANSGA-III), and Non-Dominating Sorting Genetic Algorithm-II (NSGA-II) struggle with scalability (except ANSGA-III), computational complexity, and inability to predict the required number of controllers. This thesis proposes two novel approaches to address these challenges. First, an enhanced version of NSGA-III with a repair operator-based approach (referred to as ANSGA-III) is introduced, enabling efficient CP in SD-WAN by optimizing multiple conflicting objectives simultaneously. Second, a Stochastic Computational Graph Model with Ensemble Learning (SCGMEL) is developed, overcoming scalability and computational inefficiency associated with existing methods. SCGMEL employs stochastic gradient descent with momentum, a learning rate decay, a computational graph model, a weighted sum approach, and the XGBoost algorithm for optimization and machine learning. The XGBoost predicts the number of controllers needed and a supervised classification algorithm called Learning Vector Quantization (LVQ) is used to predict the optimal locations of controllers. Additionally, this research introduces the Improved Switch Migration Decision Algorithm (ISMDA) as part of the holistic contribution. ISMDA is implemented on each controller to ensure even load distribution throughout the controllers. It functions as a plug-and-play module, periodically checking if the load surpasses a certain limit. ISMDA improves controller throughput by approximately 7.4% over CAMD and roughly 1.1% over DALB. ISMDA also outperforms DALB and CAMD with a decrease of 5.7% and 1%, respectively, in terms of controller response time. Additionally, ISMDA outperforms DALB and CAMD with a decrease of 1.7% and 5.6%, respectively, in terms of the average frequency of migrations. The established framework results in fewer switch migrations during controller load imbalance. Finally, ISMDA proves more efficient than DALB and CAMD, with an estimated 1% and 6.4% lower average packet loss, respectively. This efficiency is a result of the proposed migration efficiency strategy, allowing ISMDA to handle higher loads and reject fewer packets. Real-world experiments were conducted using the Internet Zoo topology dataset to evaluate the proposed solutions. Six objective functions, including worst-case switch-to-controller delay, load balancing, reliability, average controller-to-controller latency, maximum controller-to-controller delay, and average switch-to-controller delay, were utilized for performance evaluation. Results demonstrated that ANSGA-III outperforms existing algorithms in terms of hypervolume indicator, execution time, convergence, diversity, and scalability. SCGMEL exhibited exceptional computational efficiency, surpassing ANSGA-III, NSGA-II, and MOPSO by 99.983%, 99.985%, and 99.446% respectively. The XGBoost regression model performed significantly better in predicting the number of controllers with a mean absolute error of 1.855751 compared to 3.829268, 3.729883, and 1.883536 for KNN, linear regression, and random forest, respectively. The proposed LVQ-based classification method achieved a test accuracy of 84% and accurately predicted six of the seven controller locations. To culminate, this study presents a refined and intelligent framework designed to optimize Controller Placement (CP) within the context of SD-WAN. The proposed solutions effectively tackle the shortcomings associated with existing algorithms, addressing challenges of scalability, intelligence (including the prediction of optimal controller numbers), and computational efficiency in the pursuit of simultaneous optimization of multiple conflicting objectives. The outcomes underscore the supremacy of the suggested methodologies and underscore their potential transformative influence on SDN deployments. Notably, the findings validate the efficacy of the proposed strategies, ANSGA-III and SCGMEL, in enhancing the optimization of controller placement within SD-WAN setups. The integration of the XGBoost regression model and LVQ-based classification technique yields precise predictions for both optimal controller quantities and their respective positions. Additionally, the ISMDA algorithm emerges as a pivotal enhancement, enhancing controller throughput, mitigating packet losses, and reducing switch migration frequency—collectively contributing to elevated standards in SDN deployments
    corecore