9,680 research outputs found

    Multiobjective strategies for New Product Development in the pharmaceutical industry

    Get PDF
    New Product Development (NPD) constitutes a challenging problem in the pharmaceutical industry, due to the characteristics of the development pipeline. Formally, the NPD problem can be stated as follows: select a set of R&D projects from a pool of candidate projects in order to satisfy several criteria (economic profitability, time to market) while coping with the uncertain nature of the projects. More precisely, the recurrent key issues are to determine the projects to develop once target molecules have been identified, their order and the level of resources to assign. In this context, the proposed approach combines discrete event stochastic simulation (Monte Carlo approach) with multiobjective genetic algorithms (NSGAII type, Non-Sorted Genetic Algorithm II) to optimize the highly combinatorial portfolio management problem. In that context, Genetic Algorithms (GAs) are particularly attractive for treating this kind of problem, due to their ability to directly lead to the so-called Pareto front and to account for the combinatorial aspect. This work is illustrated with a study case involving nine interdependent new product candidates targeting three diseases. An analysis is performed for this test bench on the different pairs of criteria both for the bi- and tricriteria optimization: large portfolios cause resource queues and delays time to launch and are eliminated by the bi- and tricriteria optimization strategy. The optimization strategy is thus interesting to detect the sequence candidates. Time is an important criterion to consider simultaneously with NPV and risk criteria. The order in which drugs are released in the pipeline is of great importance as with scheduling problems

    Multiobjective strategies for New Product Development in the pharmaceutical industry

    Get PDF
    New Product Development (NPD) constitutes a challenging problem in the pharmaceutical industry, due to the characteristics of the development pipeline. Formally, the NPD problem can be stated as follows: select a set of R&D projects from a pool of candidate projects in order to satisfy several criteria (economic profitability, time to market) while coping with the uncertain nature of the projects. More precisely, the recurrent key issues are to determine the projects to develop once target molecules have been identified, their order and the level of resources to assign. In this context, the proposed approach combines discrete event stochastic simulation (Monte Carlo approach) with multiobjective genetic algorithms (NSGAII type, Non-Sorted Genetic Algorithm II) to optimize the highly combinatorial portfolio management problem. In that context, Genetic Algorithms (GAs) are particularly attractive for treating this kind of problem, due to their ability to directly lead to the so-called Pareto front and to account for the combinatorial aspect. This work is illustrated with a study case involving nine interdependent new product candidates targeting three diseases. An analysis is performed for this test bench on the different pairs of criteria both for the bi- and tricriteria optimization: large portfolios cause resource queues and delays time to launch and are eliminated by the bi- and tricriteria optimization strategy. The optimization strategy is thus interesting to detect the sequence candidates. Time is an important criterion to consider simultaneously with NPV and risk criteria. The order in which drugs are released in the pipeline is of great importance as with scheduling problems

    Sparse experimental design : an effective an efficient way discovering better genetic algorithm structures

    Get PDF
    The focus of this paper is the demonstration that sparse experimental design is a useful strategy for developing Genetic Algorithms. It is increasingly apparent from a number of reports and papers within a variety of different problem domains that the 'best' structure for a GA may be dependent upon the application. The GA structure is defined as both the types of operators and the parameters settings used during operation. The differences observed may be linked to the nature of the problem, the type of fitness function, or the depth or breadth of the problem under investigation. This paper demonstrates that advanced experimental design may be adopted to increase the understanding of the relationships between the GA structure and the problem domain, facilitating the selection of improved structures with a minimum of effort

    AI and OR in management of operations: history and trends

    Get PDF
    The last decade has seen a considerable growth in the use of Artificial Intelligence (AI) for operations management with the aim of finding solutions to problems that are increasing in complexity and scale. This paper begins by setting the context for the survey through a historical perspective of OR and AI. An extensive survey of applications of AI techniques for operations management, covering a total of over 1200 papers published from 1995 to 2004 is then presented. The survey utilizes Elsevier's ScienceDirect database as a source. Hence, the survey may not cover all the relevant journals but includes a sufficiently wide range of publications to make it representative of the research in the field. The papers are categorized into four areas of operations management: (a) design, (b) scheduling, (c) process planning and control and (d) quality, maintenance and fault diagnosis. Each of the four areas is categorized in terms of the AI techniques used: genetic algorithms, case-based reasoning, knowledge-based systems, fuzzy logic and hybrid techniques. The trends over the last decade are identified, discussed with respect to expected trends and directions for future work suggested

    Autonomous Finite Capacity Scheduling using Biological Control Principles

    Get PDF
    The vast majority of the research efforts in finite capacity scheduling over the past several years has focused on the generation of precise and almost exact measures for the working schedule presupposing complete information and a deterministic environment. During execution, however, production may be the subject of considerable variability, which may lead to frequent schedule interruptions. Production scheduling mechanisms are developed based on centralised control architecture in which all of the knowledge base and databases are modelled at the same location. This control architecture has difficulty in handling complex manufacturing systems that require knowledge and data at different locations. Adopting biological control principles refers to the process where a schedule is developed prior to the start of the processing after considering all the parameters involved at a resource involved and updated accordingly as the process executes. This research reviews the best practices in gene transcription and translation control methods and adopts these principles in the development of an autonomous finite capacity scheduling control logic aimed at reducing excessive use of manual input in planning tasks. With autonomous decision-making functionality, finite capacity scheduling will as much as practicably possible be able to respond autonomously to schedule disruptions by deployment of proactive scheduling procedures that may be used to revise or re-optimize the schedule when unexpected events occur. The novelty of this work is the ability of production resources to autonomously take decisions and the same way decisions are taken by autonomous entities in the process of gene transcription and translation. The idea has been implemented by the integration of simulation and modelling techniques with Taguchi analysis to investigate the contributions of finite capacity scheduling factors, and determination of the ‘what if’ scenarios encountered due to the existence of variability in production processes. The control logic adopts the induction rules as used in gene expression control mechanisms, studied in biological systems. Scheduling factors are identified to that effect and are investigated to find their effects on selected performance measurements for each resource in used. How they are used to deal with variability in the process is one major objective for this research as it is because of the variability that autonomous decision making becomes of interest. Although different scheduling techniques have been applied and are successful in production planning and control, the results obtained from the inclusion of the autonomous finite capacity scheduling control logic has proved that significant improvement can still be achieved

    A Multiple Criteria Genetic Algorithm Scheduling Tool for Production Scheduling in the Capital Goods Industry

    Get PDF
    Production planners usually aim to satisfy multiple objectives. This paper describes the development of a genetic algorithm tool that finds optimum trade-offs among delivery performance, resource utilisation, and workin-progress inventory. The tool was specifically developed to meet the requirements of capital goods companies that manufacture products with deep and complex product structures with components that have long and complicated routings. The model takes into account operation and assembly precedence relationships and finite capacity constraints. The tool was tested using various production problems that were obtained from a collaborating company. A series of experiments showed the tool provides a set of non-dominated solutions that enable the planner to choose an optimum trade-off according to their preferences. Previous research had optimised a single objective function. This is the first scheduling tool of its type that has simultaneously optimised delivery performance, resource utilisation and work-in-progress inventory. The quality of the schedules produced was significantly better than the approaches used by the collaborating company
    corecore