25,579 research outputs found
An integrated approach for requirement selection and scheduling in software release planning
It is essential for product software companies to decide which requirements should be included in the next release and to make an appropriate time plan of the development project. Compared to the extensive research done on requirement selection, very little research has been performed on time scheduling. In this paper, we introduce two integer linear programming models that integrate time scheduling into software release planning. Given the resource and precedence constraints, our first model provides a schedule for developing the requirements such that the project duration is minimized. Our second model combines requirement selection and scheduling, so that it not only maximizes revenues but also simultaneously calculates an on-time-delivery project schedule. Since requirement dependencies are essential for scheduling the development process, we present a more detailed analysis of these dependencies. Furthermore, we present two mechanisms that facilitate dynamic adaptation for over-estimation or under-estimation of revenues or processing time, one of which includes the Scrum methodology. Finally, several simulations based on real-life data are performed. The results of these simulations indicate that requirement dependency can significantly influence the requirement selection and the corresponding project plan. Moreover, the model for combined requirement selection and scheduling outperforms the sequential selection and scheduling approach in terms of efficiency and on-time delivery. \u
A Distributed Demand-Side Management Framework for the Smart Grid
This paper proposes a fully distributed Demand-Side Management system for
Smart Grid infrastructures, especially tailored to reduce the peak demand of
residential users. In particular, we use a dynamic pricing strategy, where
energy tariffs are function of the overall power demand of customers. We
consider two practical cases: (1) a fully distributed approach, where each
appliance decides autonomously its own scheduling, and (2) a hybrid approach,
where each user must schedule all his appliances. We analyze numerically these
two approaches, showing that they are characterized practically by the same
performance level in all the considered grid scenarios. We model the proposed
system using a non-cooperative game theoretical approach, and demonstrate that
our game is a generalized ordinal potential one under general conditions.
Furthermore, we propose a simple yet effective best response strategy that is
proved to converge in a few steps to a pure Nash Equilibrium, thus
demonstrating the robustness of the power scheduling plan obtained without any
central coordination of the operator or the customers. Numerical results,
obtained using real load profiles and appliance models, show that the
system-wide peak absorption achieved in a completely distributed fashion can be
reduced up to 55%, thus decreasing the capital expenditure (CAPEX) necessary to
meet the growing energy demand
Multiobjective strategies for New Product Development in the pharmaceutical industry
New Product Development (NPD) constitutes a challenging problem in the pharmaceutical industry, due to the characteristics of the development pipeline. Formally, the NPD problem can be stated as follows: select a set of R&D projects from a pool of candidate projects in order to satisfy several criteria (economic profitability, time to market) while coping with the uncertain nature of the projects. More precisely, the recurrent key issues are to determine the projects to develop once target molecules have been identified, their order and the level of resources to assign. In this context, the proposed approach combines discrete event stochastic simulation (Monte Carlo approach) with multiobjective genetic algorithms (NSGAII type, Non-Sorted Genetic Algorithm II) to optimize the highly combinatorial portfolio management problem. In that context, Genetic Algorithms (GAs) are particularly attractive for treating this kind of problem, due to their ability to directly lead to the so-called Pareto front and to account for the combinatorial aspect. This work is illustrated with a study case involving nine interdependent new product candidates targeting three diseases. An analysis is performed for this test bench on the different pairs of criteria both for the bi- and tricriteria optimization: large portfolios cause resource queues and delays time to launch and are eliminated by the bi- and tricriteria optimization strategy. The optimization strategy is thus interesting to detect the sequence candidates. Time is an important criterion to consider simultaneously with NPV and risk criteria. The order in which drugs are released in the pipeline is of great importance as with scheduling problems
Multiobjective strategies for New Product Development in the pharmaceutical industry
New Product Development (NPD) constitutes a challenging problem in the pharmaceutical industry, due to the characteristics of the development pipeline. Formally, the NPD problem can be stated as follows: select a set of R&D projects from a pool of candidate projects in order to satisfy several criteria (economic profitability, time to market) while coping with the uncertain nature of the projects. More precisely, the recurrent key issues are to determine the projects to develop once target molecules have been identified, their order and the level of resources to assign. In this context, the proposed approach combines discrete event stochastic simulation (Monte Carlo approach) with multiobjective genetic algorithms (NSGAII type, Non-Sorted Genetic Algorithm II) to optimize the highly combinatorial portfolio management problem. In that context, Genetic Algorithms (GAs) are particularly attractive for treating this kind of problem, due to their ability to directly lead to the so-called Pareto front and to account for the combinatorial aspect. This work is illustrated with a study case involving nine interdependent new product candidates targeting three diseases. An analysis is performed for this test bench on the different pairs of criteria both for the bi- and tricriteria optimization: large portfolios cause resource queues and delays time to launch and are eliminated by the bi- and tricriteria optimization strategy. The optimization strategy is thus interesting to detect the sequence candidates. Time is an important criterion to consider simultaneously with NPV and risk criteria. The order in which drugs are released in the pipeline is of great importance as with scheduling problems
A survey of QoS-aware web service composition techniques
Web service composition can be briefly described as the process of aggregating services with disparate functionalities into a new composite service in order to meet increasingly complex needs of users. Service composition process has been accurate on dealing with services having disparate functionalities, however, over the years the number of web services in particular that exhibit similar functionalities and varying Quality of Service (QoS) has significantly increased. As such, the problem becomes how to select appropriate web services such that the QoS of the resulting composite service is maximized or, in some cases, minimized. This constitutes an NP-hard problem as it is complicated and difficult to solve. In this paper, a discussion of concepts of web service composition and a holistic review of current service composition techniques proposed in literature is presented. Our review spans several publications in the field that can serve as a road map for future research
Paraiso : An Automated Tuning Framework for Explicit Solvers of Partial Differential Equations
We propose Paraiso, a domain specific language embedded in functional
programming language Haskell, for automated tuning of explicit solvers of
partial differential equations (PDEs) on GPUs as well as multicore CPUs. In
Paraiso, one can describe PDE solving algorithms succinctly using tensor
equations notation. Hydrodynamic properties, interpolation methods and other
building blocks are described in abstract, modular, re-usable and combinable
forms, which lets us generate versatile solvers from little set of Paraiso
source codes.
We demonstrate Paraiso by implementing a compressive hydrodynamics solver. A
single source code less than 500 lines can be used to generate solvers of
arbitrary dimensions, for both multicore CPUs and GPUs. We demonstrate both
manual annotation based tuning and evolutionary computing based automated
tuning of the program.Comment: 52 pages, 14 figures, accepted for publications in Computational
Science and Discover
Numerical product design: Springback prediction, compensation and optimization
Numerical simulations are being deployed widely for product design. However, the accuracy of the numerical tools is not yet always sufficiently accurate and reliable. This article focuses on the current state and recent developments in different stages of product design: springback prediction, springback compensation and optimization by finite element (FE) analysis. To improve the springback prediction by FE analysis, guidelines regarding the mesh discretization are provided and a new through-thickness integration scheme for shell elements is launched. In the next stage of virtual product design the product is compensated for springback. Currently, deformations due to springback are manually compensated in the industry. Here, a procedure to automatically compensate the tool geometry, including the CAD description, is presented and it is successfully applied to an industrial automotive part. The last stage in virtual product design comprises optimization. This article presents an optimization scheme which is capable of designing optimal and robust metal forming processes efficiently
An agent-based fuzzy cognitive map approach to the strategic marketing planning for industrial firms
This is the post-print version of the final paper published in Industrial Marketing Management. The published article is available from the link below. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. Copyright @ 2013 Elsevier B.V.Industrial marketing planning is a typical example of an unstructured decision making problem due to the large number of variables to consider and the uncertainty imposed on those variables. Although abundant studies identified barriers and facilitators of effective industrial marketing planning in practice, the literature still lacks practical tools and methods that marketing managers can use for the task. This paper applies fuzzy cognitive maps (FCM) to industrial marketing planning. In particular, agent based inference method is proposed to overcome dynamic relationships, time lags, and reusability issues of FCM evaluation. MACOM simulator also is developed to help marketing managers conduct what-if scenarios to see the impacts of possible changes on the variables defined in an FCM that represents industrial marketing planning problem. The simulator is applied to an industrial marketing planning problem for a global software service company in South Korea. This study has practical implication as it supports marketing managers for industrial marketing planning that has large number of variables and their cause–effect relationships. It also contributes to FCM theory by providing an agent based method for the inference of FCM. Finally, MACOM also provides academics in the industrial marketing management discipline with a tool for developing and pre-verifying a conceptual model based on qualitative knowledge of marketing practitioners.Ministry of Education, Science and Technology (Korea
- …