2,772 research outputs found

    Integrating forecasting in metaheuristic methods to solve dynamic routing problems: evidence from the logistic processes of tuna vessels

    Get PDF
    The multiple Traveling Salesman Problem (mTSP) is a widespread phenomenon in real-life scenarios, and in fact it has been addressed from multiple perspectives in recent decades. However, mTSP in dynamic circumstances entails a greater complexity that recent approaches are still trying to grasp. Beyond time windows, capacity and other parameters that characterize the dynamics of each scenario, moving targets is one of the underdeveloped issues in the field of mTSP. The approach of this paper harnesses a simple prediction method to prove that integrating forecasting within a metaheuristic evolutionary-based method, such as genetic algorithms, can yield better results in a dynamic scenario than their simple non-predictive version. Real data is used from the retrieval of Fish Aggregating Devices (FADs) by tuna vessels in the Indian Ocean. Based on historical data registered by the GPS system of the buoys attached to the devices, their trajectory is firstly forecast to feed subsequently the functioning of a genetic algorithm that searches for the optimal route of tuna vessels in terms of total distance traveled. Thus, although valid for static cases and for the Vehicle Routing Problem (VRP), the main contribution of this method over existing literature lies in its application as a global search method to solve the multiple TSP with moving targets in many dynamic real-life optimization problems.Ministerio de Economía y Competitividad | Ref. ECO2016-76625-RXunta de Galicia | Ref. GRC2014/02

    Robust optimality and duality for composite uncertain multiobjective optimization in Asplund spaces with its applications

    Full text link
    This article is devoted to investigate a nonsmooth/nonconvex uncertain multiobjective optimization problem with composition fields ((\hyperlink{CUP}{\mathrm{CUP}}) for brevity) over arbitrary Asplund spaces. Employing some advanced techniques of variational analysis and generalized differentiation, we establish necessary optimality conditions for weakly robust efficient solutions of (\hyperlink{CUP}{\mathrm{CUP}}) in terms of the limiting subdifferential. Sufficient conditions for the existence of (weakly) robust efficient solutions to such a problem are also driven under the new concept of pseudo-quasi convexity for composite functions. We formulate a Mond-Weir-type robust dual problem to the primal problem (\hyperlink{CUP}{\mathrm{CUP}}), and explore weak, strong, and converse duality properties. In addition, the obtained results are applied to an approximate uncertain multiobjective problem and a composite uncertain multiobjective problem with linear operators.Comment: arXiv admin note: substantial text overlap with arXiv:2105.14366, arXiv:2205.0114

    Vibration-based damage localisation: Impulse response identification and model updating methods

    Get PDF
    Structural health monitoring has gained more and more interest over the recent decades. As the technology has matured and monitoring systems are employed commercially, the development of more powerful and precise methods is the logical next step in this field. Especially vibration sensor networks with few measurement points combined with utilisation of ambient vibration sources are attractive for practical applications, as this approach promises to be cost-effective while requiring minimal modification to the monitored structures. Since efficient methods for damage detection have already been developed for such sensor networks, the research focus shifts towards extracting more information from the measurement data, in particular to the localisation and quantification of damage. Two main concepts have produced promising results for damage localisation. The first approach involves a mechanical model of the structure, which is used in a model updating scheme to find the damaged areas of the structure. Second, there is a purely data-driven approach, which relies on residuals of vibration estimations to find regions where damage is probable. While much research has been conducted following these two concepts, different approaches are rarely directly compared using the same data sets. Therefore, this thesis presents advanced methods for vibration-based damage localisation using model updating as well as a data-driven method and provides a direct comparison using the same vibration measurement data. The model updating approach presented in this thesis relies on multiobjective optimisation. Hence, the applied numerical optimisation algorithms are presented first. On this basis, the model updating parameterisation and objective function formulation is developed. The data-driven approach employs residuals from vibration estimations obtained using multiple-input finite impulse response filters. Both approaches are then verified using a simulated cantilever beam considering multiple damage scenarios. Finally, experimentally obtained data from an outdoor girder mast structure is used to validate the approaches. In summary, this thesis provides an assessment of model updating and residual-based damage localisation by means of verification and validation cases. It is found that the residual-based method exhibits numerical performance sufficient for real-time applications while providing a high sensitivity towards damage. However, the localisation accuracy is found to be superior using the model updating method

    A new approach to portfolio selection based on forecasting

    Get PDF
    In this paper we analyze the portfolio selection problem from a novel perspective based on the analysis and prediction of the time series corresponding to the portfolio’s value. Namely, we define the value of a particular portfolio at the time of its acquisition. Using the time series of historical prices of the different financial assets, we calculate backward the value that said portfolio would have had in past time periods. A damped trend model is then used to analyze this time series and to predict the future values of the portfolio, providing estimates of the mean and variance for different forecasting horizons. These measures are used to formulate the portfolio selection problem, which is solved using a multi-objective genetic algorithm. To show the performance of this procedure, we use a data set of asset prices from the New York Stock Market

    A new differential evolution using a bilevel optimization model for solving generalized multi-point dynamic aggregation problems

    Get PDF
    The multi-point dynamic aggregation problem (MPDAP) comes mainly from real-world applications, which is characterized by dynamic task assignation and routing optimization with limited resources. Due to the dynamic allocation of tasks, more than one optimization objective, limited resources, and other factors involved, the computational complexity of both route programming and resource allocation optimization is a growing problem. In this manuscript, a task scheduling problem of fire-fighting robots is investigated and solved, and serves as a representative multi-point dynamic aggregation problem. First, in terms of two optimized objectives, the cost and completion time, a new bilevel programming model is presented, in which the task cost is taken as the leader's objective. In addition, in order to effectively solve the bilevel model, a differential evolution is developed based on a new matrix coding scheme. Moreover, some percentage of high-quality solutions are applied in mutation and selection operations, which helps to generate potentially better solutions and keep them into the next generation of population. Finally, the experimental results show that the proposed algorithm is feasible and effective in dealing with the multi-point dynamic aggregation problem

    Multi-Objective Density Diagrams Developed With Machine Learning Models to Optimize Sustainability and Cost-Efficiency of UHPC Mix Design

    Get PDF
    The emergence of ultra-high-performance concrete (UHPC) as an attractive solution for precast and prestressed applications has coincided with global efforts towards sustainable construction. The increasing need for tools capable of intuitively demonstrating the effect of concrete mixture composition on mechanical performance, cost and eco-efficiency concurrently has motivated this work in an effort to promote design of more sustainable solutions to help meet environmental goals. Such tools are needed to effectively evaluate the environmental impact of UHPC given the outstanding mechanical properties of the material coupled with high volumetric embodied CO2. Meanwhile, artificial intelligence (AI) techniques have emerged as a great opportunity for game-changing tools capable of effectively modeling the synergistic relationships between mix proportions and material performance. This work couples machine learning models with orthogonal arrays to generate machine-learning-based tools to evaluate the tradeoffs between emissions, cost and mechanical performance concurrently. Random forest and k-nearest neighbors’ models are ensembled to predict the compressive strength of UHPC mixtures and generate Performance Density Diagrams (PDDs). These predicted strengths are then coupled with volumetric environmental factors and unit costs to generate eco- and cost-efficiency density diagrams. The makeup of these tools facilitates the evaluation of rather complicated trends associated with mix proportions and multi-objective outcomes, allowing AI-based tools to be of easy use by industry personnel on a daily basis, while serving as decision-making aids during mix design stages and provide proof of mixture optimization that could be introduced in Environmental Product Declarations. The PDD developed herein enabled the design of a mix with compressive strength of 155 MPa, while keeping the aggregate-to-cementitious ratio above unit. Other mixtures were developed from these models and compared to several different concretes from the literature. Results show that high paste content, high strength (and ultra-high strength) concrete technologies are not necessarily detrimental to cost or eco efficiencies. For the different indices evaluated, optimum solutions were mostly obtained with these types of concrete, which means that industry trends toward requiring minimization of embodied CO2 in concrete on a per volume basis are misguided and do not minimize the embodied CO2 in concrete structures

    Undergraduate Catalog of Studies, 2022-2023

    Get PDF

    Serial-batch scheduling – the special case of laser-cutting machines

    Get PDF
    The dissertation deals with a problem in the field of short-term production planning, namely the scheduling of laser-cutting machines. The object of decision is the grouping of production orders (batching) and the sequencing of these order groups on one or more machines (scheduling). This problem is also known in the literature as "batch scheduling problem" and belongs to the class of combinatorial optimization problems due to the interdependencies between the batching and the scheduling decisions. The concepts and methods used are mainly from production planning, operations research and machine learning

    Subgroup discovery for structured target concepts

    Get PDF
    The main object of study in this thesis is subgroup discovery, a theoretical framework for finding subgroups in data—i.e., named sub-populations— whose behaviour with respect to a specified target concept is exceptional when compared to the rest of the dataset. This is a powerful tool that conveys crucial information to a human audience, but despite past advances has been limited to simple target concepts. In this work we propose algorithms that bring this framework to novel application domains. We introduce the concept of representative subgroups, which we use not only to ensure the fairness of a sub-population with regard to a sensitive trait, such as race or gender, but also to go beyond known trends in the data. For entities with additional relational information that can be encoded as a graph, we introduce a novel measure of robust connectedness which improves on established alternative measures of density; we then provide a method that uses this measure to discover which named sub-populations are more well-connected. Our contributions within subgroup discovery crescent with the introduction of kernelised subgroup discovery: a novel framework that enables the discovery of subgroups on i.i.d. target concepts with virtually any kind of structure. Importantly, our framework additionally provides a concrete and efficient tool that works out-of-the-box without any modification, apart from specifying the Gramian of a positive definite kernel. To use within kernelised subgroup discovery, but also on any other kind of kernel method, we additionally introduce a novel random walk graph kernel. Our kernel allows the fine tuning of the alignment between the vertices of the two compared graphs, during the count of the random walks, while we also propose meaningful structure-aware vertex labels to utilise this new capability. With these contributions we thoroughly extend the applicability of subgroup discovery and ultimately re-define it as a kernel method.Der Hauptgegenstand dieser Arbeit ist die Subgruppenentdeckung (Subgroup Discovery), ein theoretischer Rahmen für das Auffinden von Subgruppen in Daten—d. h. benannte Teilpopulationen—deren Verhalten in Bezug auf ein bestimmtes Targetkonzept im Vergleich zum Rest des Datensatzes außergewöhnlich ist. Es handelt sich hierbei um ein leistungsfähiges Instrument, das einem menschlichen Publikum wichtige Informationen vermittelt. Allerdings ist es trotz bisherigen Fortschritte auf einfache Targetkonzepte beschränkt. In dieser Arbeit schlagen wir Algorithmen vor, die diesen Rahmen auf neuartige Anwendungsbereiche übertragen. Wir führen das Konzept der repräsentativen Untergruppen ein, mit dem wir nicht nur die Fairness einer Teilpopulation in Bezug auf ein sensibles Merkmal wie Rasse oder Geschlecht sicherstellen, sondern auch über bekannte Trends in den Daten hinausgehen können. Für Entitäten mit zusätzlicher relationalen Information, die als Graph kodiert werden kann, führen wir ein neuartiges Maß für robuste Verbundenheit ein, das die etablierten alternativen Dichtemaße verbessert; anschließend stellen wir eine Methode bereit, die dieses Maß verwendet, um herauszufinden, welche benannte Teilpopulationen besser verbunden sind. Unsere Beiträge in diesem Rahmen gipfeln in der Einführung der kernelisierten Subgruppenentdeckung: ein neuartiger Rahmen, der die Entdeckung von Subgruppen für u.i.v. Targetkonzepten mit praktisch jeder Art von Struktur ermöglicht. Wichtigerweise, unser Rahmen bereitstellt zusätzlich ein konkretes und effizientes Werkzeug, das ohne jegliche Modifikation funktioniert, abgesehen von der Angabe des Gramian eines positiv definitiven Kernels. Für den Einsatz innerhalb der kernelisierten Subgruppentdeckung, aber auch für jede andere Art von Kernel-Methode, führen wir zusätzlich einen neuartigen Random-Walk-Graph-Kernel ein. Unser Kernel ermöglicht die Feinabstimmung der Ausrichtung zwischen den Eckpunkten der beiden unter-Vergleich-gestelltenen Graphen während der Zählung der Random Walks, während wir auch sinnvolle strukturbewusste Vertex-Labels vorschlagen, um diese neue Fähigkeit zu nutzen. Mit diesen Beiträgen erweitern wir die Anwendbarkeit der Subgruppentdeckung gründlich und definieren wir sie im Endeffekt als Kernel-Methode neu

    Advances and Applications of DSmT for Information Fusion. Collected Works, Volume 5

    Get PDF
    This fifth volume on Advances and Applications of DSmT for Information Fusion collects theoretical and applied contributions of researchers working in different fields of applications and in mathematics, and is available in open-access. The collected contributions of this volume have either been published or presented after disseminating the fourth volume in 2015 in international conferences, seminars, workshops and journals, or they are new. The contributions of each part of this volume are chronologically ordered. First Part of this book presents some theoretical advances on DSmT, dealing mainly with modified Proportional Conflict Redistribution Rules (PCR) of combination with degree of intersection, coarsening techniques, interval calculus for PCR thanks to set inversion via interval analysis (SIVIA), rough set classifiers, canonical decomposition of dichotomous belief functions, fast PCR fusion, fast inter-criteria analysis with PCR, and improved PCR5 and PCR6 rules preserving the (quasi-)neutrality of (quasi-)vacuous belief assignment in the fusion of sources of evidence with their Matlab codes. Because more applications of DSmT have emerged in the past years since the apparition of the fourth book of DSmT in 2015, the second part of this volume is about selected applications of DSmT mainly in building change detection, object recognition, quality of data association in tracking, perception in robotics, risk assessment for torrent protection and multi-criteria decision-making, multi-modal image fusion, coarsening techniques, recommender system, levee characterization and assessment, human heading perception, trust assessment, robotics, biometrics, failure detection, GPS systems, inter-criteria analysis, group decision, human activity recognition, storm prediction, data association for autonomous vehicles, identification of maritime vessels, fusion of support vector machines (SVM), Silx-Furtif RUST code library for information fusion including PCR rules, and network for ship classification. Finally, the third part presents interesting contributions related to belief functions in general published or presented along the years since 2015. These contributions are related with decision-making under uncertainty, belief approximations, probability transformations, new distances between belief functions, non-classical multi-criteria decision-making problems with belief functions, generalization of Bayes theorem, image processing, data association, entropy and cross-entropy measures, fuzzy evidence numbers, negator of belief mass, human activity recognition, information fusion for breast cancer therapy, imbalanced data classification, and hybrid techniques mixing deep learning with belief functions as well
    corecore