7,375 research outputs found
Multi-Information Source Fusion and Optimization to Realize ICME: Application to Dual Phase Materials
Integrated Computational Materials Engineering (ICME) calls for the
integration of computational tools into the materials and parts development
cycle, while the Materials Genome Initiative (MGI) calls for the acceleration
of the materials development cycle through the combination of experiments,
simulation, and data. As they stand, both ICME and MGI do not prescribe how to
achieve the necessary tool integration or how to efficiently exploit the
computational tools, in combination with experiments, to accelerate the
development of new materials and materials systems. This paper addresses the
first issue by putting forward a framework for the fusion of information that
exploits correlations among sources/models and between the sources and `ground
truth'. The second issue is addressed through a multi-information source
optimization framework that identifies, given current knowledge, the next best
information source to query and where in the input space to query it via a
novel value-gradient policy. The querying decision takes into account the
ability to learn correlations between information sources, the resource cost of
querying an information source, and what a query is expected to provide in
terms of improvement over the current state. The framework is demonstrated on
the optimization of a dual-phase steel to maximize its strength-normalized
strain hardening rate. The ground truth is represented by a
microstructure-based finite element model while three low fidelity information
sources---i.e. reduced order models---based on different homogenization
assumptions---isostrain, isostress and isowork---are used to efficiently and
optimally query the materials design space.Comment: 19 pages, 11 figures, 5 table
Constrained multi-objective optimization of process design parameters in settings with scarce data: an application to adhesive bonding
Adhesive joints are increasingly used in industry for a wide variety of
applications because of their favorable characteristics such as high
strength-to-weight ratio, design flexibility, limited stress concentrations,
planar force transfer, good damage tolerance and fatigue resistance. Finding
the optimal process parameters for an adhesive bonding process is challenging:
the optimization is inherently multi-objective (aiming to maximize break
strength while minimizing cost) and constrained (the process should not result
in any visual damage to the materials, and stress tests should not result in
failures that are adhesion-related). Real life physical experiments in the lab
are expensive to perform; traditional evolutionary approaches (such as genetic
algorithms) are then ill-suited to solve the problem, due to the prohibitive
amount of experiments required for evaluation. In this research, we
successfully applied specific machine learning techniques (Gaussian Process
Regression and Logistic Regression) to emulate the objective and constraint
functions based on a limited amount of experimental data. The techniques are
embedded in a Bayesian optimization algorithm, which succeeds in detecting
Pareto-optimal process settings in a highly efficient way (i.e., requiring a
limited number of extra experiments)
A survey on handling computationally expensive multiobjective optimization problems with evolutionary algorithms
This is the author accepted manuscript. The final version is available from Springer Verlag via the DOI in this record.Evolutionary algorithms are widely used for solving multiobjective optimization problems but are often criticized because of a large number of function evaluations needed. Approximations, especially function approximations, also referred to as surrogates or metamodels are commonly used in the literature to reduce the computation time. This paper presents a survey of 45 different recent algorithms proposed in the literature between 2008 and 2016 to handle computationally expensive multiobjective optimization problems. Several algorithms are discussed based on what kind of an approximation such as problem, function or fitness approximation they use. Most emphasis is given to function approximation-based algorithms. We also compare these algorithms based on different criteria such as metamodeling technique and evolutionary algorithm used, type and dimensions of the problem solved, handling constraints, training time and the type of evolution control. Furthermore, we identify and discuss some promising elements and major issues among algorithms in the literature related to using an approximation and numerical settings used. In addition, we discuss selecting an algorithm to solve a given computationally expensive multiobjective optimization problem based on the dimensions in both objective and decision spaces and the computation budget available.The research of Tinkle Chugh was funded by the COMAS Doctoral Program (at the University of Jyväskylä) and FiDiPro Project DeCoMo (funded by Tekes, the Finnish Funding Agency for Innovation), and the research of Dr. Karthik Sindhya was funded by SIMPRO project funded by Tekes as well as DeCoMo
Towards an evolvable cancer treatment simulator
© 2019 Elsevier B.V. The use of high-fidelity computational simulations promises to enable high-throughput hypothesis testing and optimisation of cancer therapies. However, increasing realism comes at the cost of increasing computational requirements. This article explores the use of surrogate-assisted evolutionary algorithms to optimise the targeted delivery of a therapeutic compound to cancerous tumour cells with the multicellular simulator, PhysiCell. The use of both Gaussian process models and multi-layer perceptron neural network surrogate models are investigated. We find that evolutionary algorithms are able to effectively explore the parameter space of biophysical properties within the agent-based simulations, minimising the resulting number of cancerous cells after a period of simulated treatment. Both model-assisted algorithms are found to outperform a standard evolutionary algorithm, demonstrating their ability to perform a more effective search within the very small evaluation budget. This represents the first use of efficient evolutionary algorithms within a high-throughput multicellular computing approach to find therapeutic design optima that maximise tumour regression
Solving optimisation problems in metal forming using FEM: A metamodel based optimisation algorithm
During the last decades, Finite Element (FEM) simulations of metal forming processes have\ud
become important tools for designing feasible production processes. In more recent years,\ud
several authors recognised the potential of coupling FEM simulations to mathematical opti-\ud
misation algorithms to design optimal metal forming processes instead of only feasible ones.\ud
This report describes the selection, development and implementation of an optimisa-\ud
tion algorithm for solving optimisation problems for metal forming processes using time\ud
consuming FEM simulations. A Sequential Approximate Optimisation algorithm is pro-\ud
posed, which incorporates metamodelling techniques and sequential improvement strate-\ud
gies for enhancing the e±ciency of the algorithm. The algorithm has been implemented in\ud
MATLABr and can be used in combination with any Finite Element code for simulating\ud
metal forming processes.\ud
The good applicability of the proposed optimisation algorithm within the ¯eld of metal\ud
forming has been demonstrated by applying it to optimise the internal pressure and ax-\ud
ial feeding load paths for manufacturing a simple hydroformed product. Resulting was\ud
a constantly distributed wall thickness throughout the ¯nal product. Subsequently, the\ud
algorithm was compared to other optimisation algorithms for optimising metal forming\ud
by applying it to two more complicated forging examples. In both cases, the geometry of\ud
the preform was optimised. For one forging application, the algorithm managed to solve\ud
a folding defect. For the other application both the folding susceptibility and the energy\ud
consumption required for forging the part were reduced by 10% w.r.t. the forging process\ud
proposed by the forging company. The algorithm proposed in this report yielded better\ud
results than the optimisation algorithms it was compared to
Fold Family-Regularized Bayesian Optimization for Directed Protein Evolution
Directed Evolution (DE) is a technique for protein engineering that involves iterative rounds of mutagenesis and screening to search for sequences that optimize a given property (ex. binding affinity to a specified target). Unfortunately, the underlying optimization problem is under-determined, and so mutations introduced to improve the specified property may come at the expense of unmeasured, but nevertheless important properties (ex. subcellular localization). We seek to address this issue by incorporating a fold-specific regularization factor into the optimization problem. The regularization factor biases the search towards designs that resemble sequences from the fold family to which the protein belongs. We applied our method to a large library of protein GB1 mutants with binding affinity measurements to IgG-Fc. Our results demonstrate that the regularized optimization problem produces more native-like GB1 sequences with only a minor decrease in binding affinity. Specifically, the log-odds of our designs under a generative model of the GB1 fold family are between 41-45% higher than those obtained without regularization, with only a 7% drop in binding affinity. Thus, our method is capable of making a trade-off between competing traits. Moreover, we demonstrate that our active-learning driven approach reduces the wet-lab burden to identify optimal GB1 designs by 67%, relative to recent results from the Arnold lab on the same data
- …