7,505 research outputs found
Recommended from our members
Bicriteria scheduling of a two-machine flowshop with sequence-dependent setup times
The official published version of the article can be found at the link below.A two-machine flowshop scheduling problem is addressed to minimize setups and makespan where each job is characterized by a pair of attributes that entail setups on each machine. The setup times are sequence-dependent on both machines. It is shown that these objectives conflict, so the Pareto optimization approach is considered. The scheduling problems considering either of these objectives are NP-hard , so exact optimization techniques are impractical for large-sized problems. We propose two multi-objective metaheurisctics based on genetic algorithms (MOGA) and simulated annealing (MOSA) to find approximations of Pareto-optimal sets. The performances of these approaches are compared with lower bounds for small problems. In larger problems, performance of the proposed algorithms are compared with each other. Experimentations revealed that both algorithms perform very similar on small problems. Moreover, it was observed that MOGA outperforms MOSA in terms of the quality of solutions on larger problems.Partial Funding from EPSRC under grant EP/D050863/1
Multiobjective Tactical Planning under Uncertainty for Air Traffic Flow and Capacity Management
We investigate a method to deal with congestion of sectors and delays in the
tactical phase of air traffic flow and capacity management. It relies on
temporal objectives given for every point of the flight plans and shared among
the controllers in order to create a collaborative environment. This would
enhance the transition from the network view of the flow management to the
local view of air traffic control. Uncertainty is modeled at the trajectory
level with temporal information on the boundary points of the crossed sectors
and then, we infer the probabilistic occupancy count. Therefore, we can model
the accuracy of the trajectory prediction in the optimization process in order
to fix some safety margins. On the one hand, more accurate is our prediction;
more efficient will be the proposed solutions, because of the tighter safety
margins. On the other hand, when uncertainty is not negligible, the proposed
solutions will be more robust to disruptions. Furthermore, a multiobjective
algorithm is used to find the tradeoff between the delays and congestion, which
are antagonist in airspace with high traffic density. The flow management
position can choose manually, or automatically with a preference-based
algorithm, the adequate solution. This method is tested against two instances,
one with 10 flights and 5 sectors and one with 300 flights and 16 sectors.Comment: IEEE Congress on Evolutionary Computation (2013). arXiv admin note:
substantial text overlap with arXiv:1309.391
A Study of Archiving Strategies in Multi-Objective PSO for Molecular Docking
Molecular docking is a complex optimization problem aimed at predicting the position of a ligand molecule in the active site of a receptor with the lowest binding energy. This problem can be formulated as a bi-objective optimization problem by minimizing the binding energy and the Root Mean Square Deviation (RMSD) difference in the coordinates of ligands. In this context, the SMPSO multi-objective swarm-intelligence algorithm has shown a remarkable performance. SMPSO is characterized by having an external archive used to store the non-dominated solutions and also as the basis of the leader selection strategy. In this paper, we analyze several SMPSO variants based on different archiving strategies in the scope of a benchmark of molecular docking instances. Our study reveals that the SMPSOhv, which uses an hypervolume contribution based archive, shows the overall best performance.Universidad de Málaga. Campus de Excelencia Internacional Andalucía Tech
Energy-Efficient selective activation in Femtocell Networks
Provisioning the capacity of wireless networks is difficult when peak load is significantly higher than average load, for example, in public spaces like airports or train stations. Service providers can use femtocells and small cells to increase local capacity, but deploying enough femtocells to serve peak loads requires a large number of femtocells that will remain idle most of the time, which wastes a significant amount of power.
To reduce the energy consumption of over-provisioned femtocell networks, we formulate a femtocell selective activation problem, which we formalize as an integer nonlinear optimization problem. Then we introduce GREENFEMTO, a distributed femtocell selective activation algorithm that deactivates idle femtocells to
save power and activates them on-the-fly as the number of users increases. We prove that GREENFEMTO converges to a locally Pareto optimal solution and demonstrate its performance using extensive simulations of an LTE wireless system. Overall, we find that GREENFEMTO requires up to 55% fewer femtocells to serve a given user load, relative to an existing femtocell power-saving procedure, and comes within 15% of a globally optimal solution
Numerical Investigation and Optimization of a Flushwall Injector for Scramjet Applications at Hypervelocity Flow Conditions
An investigation utilizing Reynolds-averaged simulations (RAS) was performed in order to demonstrate the use of design and analysis of computer experiments (DACE) methods in Sandias DAKOTA software package for surrogate modeling and optimization. These methods were applied to a flow- path fueled with an interdigitated flushwall injector suitable for scramjet applications at hyper- velocity conditions and ascending along a constant dynamic pressure flight trajectory. The flight Mach number, duct height, spanwise width, and injection angle were the design variables selected to maximize two objective functions: the thrust potential and combustion efficiency. Because the RAS of this case are computationally expensive, surrogate models are used for optimization. To build a surrogate model a RAS database is created. The sequence of the design variables comprising the database were generated using a Latin hypercube sampling (LHS) method. A methodology was also developed to automatically build geometries and generate structured grids for each design point. The ensuing RAS analysis generated the simulation database from which the two objective functions were computed using a one-dimensionalization (1D) of the three-dimensional simulation data. The data were fitted using four surrogate models: an artificial neural network (ANN), a cubic polynomial, a quadratic polynomial, and a Kriging model. Variance-based decomposition showed that both objective functions were primarily driven by changes in the duct height. Multiobjective design optimization was performed for all four surrogate models via a genetic algorithm method. Optimal solutions were obtained at the upper and lower bounds of the flight Mach number range. The Kriging model predicted an optimal solution set that exhibited high values for both objective functions. Additionally, three challenge points were selected to assess the designs on the Pareto fronts. Further sampling among the designs of the Pareto fronts may be required to lower the surrogate model errors and perform more accurate surrogate-model-based optimization
Approximating multivariate posterior distribution functions from Monte Carlo samples for sequential Bayesian inference
An important feature of Bayesian statistics is the opportunity to do
sequential inference: the posterior distribution obtained after seeing a
dataset can be used as prior for a second inference. However, when Monte Carlo
sampling methods are used for inference, we only have a set of samples from the
posterior distribution. To do sequential inference, we then either have to
evaluate the second posterior at only these locations and reweight the samples
accordingly, or we can estimate a functional description of the posterior
probability distribution from the samples and use that as prior for the second
inference. Here, we investigated to what extent we can obtain an accurate joint
posterior from two datasets if the inference is done sequentially rather than
jointly, under the condition that each inference step is done using Monte Carlo
sampling. To test this, we evaluated the accuracy of kernel density estimates,
Gaussian mixtures, vine copulas and Gaussian processes in approximating
posterior distributions, and then tested whether these approximations can be
used in sequential inference. In low dimensionality, Gaussian processes are
more accurate, whereas in higher dimensionality Gaussian mixtures or vine
copulas perform better. In our test cases, posterior approximations are
preferable over direct sample reweighting, although joint inference is still
preferable over sequential inference. Since the performance is case-specific,
we provide an R package mvdens with a unified interface for the density
approximation methods
- …