4,375 research outputs found

    A Comparative Study of Efficient Initialization Methods for the K-Means Clustering Algorithm

    Full text link
    K-means is undoubtedly the most widely used partitional clustering algorithm. Unfortunately, due to its gradient descent nature, this algorithm is highly sensitive to the initial placement of the cluster centers. Numerous initialization methods have been proposed to address this problem. In this paper, we first present an overview of these methods with an emphasis on their computational efficiency. We then compare eight commonly used linear time complexity initialization methods on a large and diverse collection of data sets using various performance criteria. Finally, we analyze the experimental results using non-parametric statistical tests and provide recommendations for practitioners. We demonstrate that popular initialization methods often perform poorly and that there are in fact strong alternatives to these methods.Comment: 17 pages, 1 figure, 7 table

    Holistic, data-driven, service and supply chain optimisation: linked optimisation.

    Get PDF
    The intensity of competition and technological advancements in the business environment has made companies collaborate and cooperate together as a means of survival. This creates a chain of companies and business components with unified business objectives. However, managing the decision-making process (like scheduling, ordering, delivering and allocating) at the various business components and maintaining a holistic objective is a huge business challenge, as these operations are complex and dynamic. This is because the overall chain of business processes is widely distributed across all the supply chain participants; therefore, no individual collaborator has a complete overview of the processes. Increasingly, such decisions are automated and are strongly supported by optimisation algorithms - manufacturing optimisation, B2B ordering, financial trading, transportation scheduling and allocation. However, most of these algorithms do not incorporate the complexity associated with interacting decision-making systems like supply chains. It is well-known that decisions made at one point in supply chains can have significant consequences that ripple through linked production and transportation systems. Recently, global shocks to supply chains (COVID-19, climate change, blockage of the Suez Canal) have demonstrated the importance of these interdependencies, and the need to create supply chains that are more resilient and have significantly reduced impact on the environment. Such interacting decision-making systems need to be considered through an optimisation process. However, the interactions between such decision-making systems are not modelled. We therefore believe that modelling such interactions is an opportunity to provide computational extensions to current optimisation paradigms. This research study aims to develop a general framework for formulating and solving holistic, data-driven optimisation problems in service and supply chains. This research achieved this aim and contributes to scholarship by firstly considering the complexities of supply chain problems from a linked problem perspective. This leads to developing a formalism for characterising linked optimisation problems as a model for supply chains. Secondly, the research adopts a method for creating a linked optimisation problem benchmark by linking existing classical benchmark sets. This involves using a mix of classical optimisation problems, typically relating to supply chain decision problems, to describe different modes of linkages in linked optimisation problems. Thirdly, several techniques for linking supply chain fragmented data have been proposed in the literature to identify data relationships. Therefore, this thesis explores some of these techniques and combines them in specific ways to improve the data discovery process. Lastly, many state-of-the-art algorithms have been explored in the literature and these algorithms have been used to tackle problems relating to supply chain problems. This research therefore investigates the resilient state-of-the-art optimisation algorithms presented in the literature, and then designs suitable algorithmic approaches inspired by the existing algorithms and the nature of problem linkages to address different problem linkages in supply chains. Considering research findings and future perspectives, the study demonstrates the suitability of algorithms to different linked structures involving two sub-problems, which suggests further investigations on issues like the suitability of algorithms on more complex structures, benchmark methodologies, holistic goals and evaluation, processmining, game theory and dependency analysis

    A survey on handling computationally expensive multiobjective optimization problems with evolutionary algorithms

    Get PDF
    This is the author accepted manuscript. The final version is available from Springer Verlag via the DOI in this record.Evolutionary algorithms are widely used for solving multiobjective optimization problems but are often criticized because of a large number of function evaluations needed. Approximations, especially function approximations, also referred to as surrogates or metamodels are commonly used in the literature to reduce the computation time. This paper presents a survey of 45 different recent algorithms proposed in the literature between 2008 and 2016 to handle computationally expensive multiobjective optimization problems. Several algorithms are discussed based on what kind of an approximation such as problem, function or fitness approximation they use. Most emphasis is given to function approximation-based algorithms. We also compare these algorithms based on different criteria such as metamodeling technique and evolutionary algorithm used, type and dimensions of the problem solved, handling constraints, training time and the type of evolution control. Furthermore, we identify and discuss some promising elements and major issues among algorithms in the literature related to using an approximation and numerical settings used. In addition, we discuss selecting an algorithm to solve a given computationally expensive multiobjective optimization problem based on the dimensions in both objective and decision spaces and the computation budget available.The research of Tinkle Chugh was funded by the COMAS Doctoral Program (at the University of Jyväskylä) and FiDiPro Project DeCoMo (funded by Tekes, the Finnish Funding Agency for Innovation), and the research of Dr. Karthik Sindhya was funded by SIMPRO project funded by Tekes as well as DeCoMo

    Parallel Multi-Objective Evolutionary Algorithms: A Comprehensive Survey

    Get PDF
    Multi-Objective Evolutionary Algorithms (MOEAs) are powerful search techniques that have been extensively used to solve difficult problems in a wide variety of disciplines. However, they can be very demanding in terms of computational resources. Parallel implementations of MOEAs (pMOEAs) provide considerable gains regarding performance and scalability and, therefore, their relevance in tackling computationally expensive applications. This paper presents a survey of pMOEAs, describing a refined taxonomy, an up-to-date review of methods and the key contributions to the field. Furthermore, some of the open questions that require further research are also briefly discussed

    Metaheuristic design of feedforward neural networks: a review of two decades of research

    Get PDF
    Over the past two decades, the feedforward neural network (FNN) optimization has been a key interest among the researchers and practitioners of multiple disciplines. The FNN optimization is often viewed from the various perspectives: the optimization of weights, network architecture, activation nodes, learning parameters, learning environment, etc. Researchers adopted such different viewpoints mainly to improve the FNN's generalization ability. The gradient-descent algorithm such as backpropagation has been widely applied to optimize the FNNs. Its success is evident from the FNN's application to numerous real-world problems. However, due to the limitations of the gradient-based optimization methods, the metaheuristic algorithms including the evolutionary algorithms, swarm intelligence, etc., are still being widely explored by the researchers aiming to obtain generalized FNN for a given problem. This article attempts to summarize a broad spectrum of FNN optimization methodologies including conventional and metaheuristic approaches. This article also tries to connect various research directions emerged out of the FNN optimization practices, such as evolving neural network (NN), cooperative coevolution NN, complex-valued NN, deep learning, extreme learning machine, quantum NN, etc. Additionally, it provides interesting research challenges for future research to cope-up with the present information processing era

    Machine learning techniques focusing on the energy performance of buildings: A dimensions and methods analysis

    Get PDF
    The problem of energy consumption and the importance of improving existing buildings’ energy performance are notorious. This work aims to contribute to this improvement by identifying the latest and most appropriate machine learning or statistical techniques, which analyze this problem by looking at large quantities of building energy performance certification data and other data sources. PRISMA, a well-established systematic literature review and meta-analysis method, was used to detect specific factors that influence the energy performance of buildings, resulting in an analysis of 35 papers published between 2016 and April 2021, creating a baseline for further inquiry. Through this systematic literature review and bibliometric analysis, machine learning and statistical approaches primarily based on building energy certification data were identified and analyzed in two groups: (1) automatic evaluation of buildings’ energy performance and, (2) prediction of energy-efficient retrofit measures. The main contribution of our study is a conceptual and theoretical framework applicable in the analysis of the energy performance of buildings with intelligent computational methods. With our framework, the reader can understand which approaches are most used and more appropriate for analyzing the energy performance of different types of buildings, discussing the dimensions that are better used in such approaches.info:eu-repo/semantics/publishedVersio

    Machine Learning Techniques Focusing on the Energy Performance of Buildings: A Dimensions and Methods Analysis

    Get PDF
    109 “Consumo SMART”. This work is partially funded by national funds through FCT—Foundation for Science and Technology, I.P., under the project FCT UIDB/04466/2020.The problem of energy consumption and the importance of improving existing buildings’ energy performance are notorious. This work aims to contribute to this improvement by identifying the latest and most appropriate machine learning or statistical techniques, which analyze this problem by looking at large quantities of building energy performance certification data and other data sources. PRISMA, a well-established systematic literature review and meta-analysis method, was used to detect specific factors that influence the energy performance of buildings, resulting in an analysis of 35 papers published between 2016 and April 2021, creating a baseline for further inquiry. Through this systematic literature review and bibliometric analysis, machine learning and statistical approaches primarily based on building energy certification data were identified and analyzed in two groups: (1) automatic evaluation of buildings’ energy performance and, (2) prediction of energy-efficient retrofit measures. The main contribution of our study is a conceptual and theoretical framework applicable in the analysis of the energy performance of buildings with intelligent computational methods. With our framework, the reader can understand which approaches are most used and more appropriate for analyzing the energy performance of different types of buildings, discussing the dimensions that are better used in such approaches.publishersversionpublishe

    Performance of a sequential versus holistic building design approach using multi-objective optimization

    Get PDF
    Integrated design processes are currently pushed forward in order to achieve net-zero energy building designs at affordable cost. Through a case study of a residential building, this paper compares a sequential versus a holistic design approach based on multi-objective optimization. In the holistic approach, 39 design variables related to the architecture and HVAC systems are simultaneously optimized. In the sequential approach, the architecture variables are first optimized; several optimal solutions are then selected for the second phase optimization of the heating system parameters. Carbon footprint, life cycle cost and thermal comfort are optimized by the algorithm NSGA-II. With only 100 computational hours, the holistic approach found 59% of the optimal solutions, whereas it took 765 h to find 41% of the optimal solutions with the sequential approach. This comparison shows the negative effects of making irreversible variable selections in the early phase of a design process, as it reduces the ability to find optimal solutions in the end
    corecore