305 research outputs found

    Multiobjective strategies for New Product Development in the pharmaceutical industry

    Get PDF
    New Product Development (NPD) constitutes a challenging problem in the pharmaceutical industry, due to the characteristics of the development pipeline. Formally, the NPD problem can be stated as follows: select a set of R&D projects from a pool of candidate projects in order to satisfy several criteria (economic profitability, time to market) while coping with the uncertain nature of the projects. More precisely, the recurrent key issues are to determine the projects to develop once target molecules have been identified, their order and the level of resources to assign. In this context, the proposed approach combines discrete event stochastic simulation (Monte Carlo approach) with multiobjective genetic algorithms (NSGAII type, Non-Sorted Genetic Algorithm II) to optimize the highly combinatorial portfolio management problem. In that context, Genetic Algorithms (GAs) are particularly attractive for treating this kind of problem, due to their ability to directly lead to the so-called Pareto front and to account for the combinatorial aspect. This work is illustrated with a study case involving nine interdependent new product candidates targeting three diseases. An analysis is performed for this test bench on the different pairs of criteria both for the bi- and tricriteria optimization: large portfolios cause resource queues and delays time to launch and are eliminated by the bi- and tricriteria optimization strategy. The optimization strategy is thus interesting to detect the sequence candidates. Time is an important criterion to consider simultaneously with NPV and risk criteria. The order in which drugs are released in the pipeline is of great importance as with scheduling problems

    Multiobjective strategies for New Product Development in the pharmaceutical industry

    Get PDF
    New Product Development (NPD) constitutes a challenging problem in the pharmaceutical industry, due to the characteristics of the development pipeline. Formally, the NPD problem can be stated as follows: select a set of R&D projects from a pool of candidate projects in order to satisfy several criteria (economic profitability, time to market) while coping with the uncertain nature of the projects. More precisely, the recurrent key issues are to determine the projects to develop once target molecules have been identified, their order and the level of resources to assign. In this context, the proposed approach combines discrete event stochastic simulation (Monte Carlo approach) with multiobjective genetic algorithms (NSGAII type, Non-Sorted Genetic Algorithm II) to optimize the highly combinatorial portfolio management problem. In that context, Genetic Algorithms (GAs) are particularly attractive for treating this kind of problem, due to their ability to directly lead to the so-called Pareto front and to account for the combinatorial aspect. This work is illustrated with a study case involving nine interdependent new product candidates targeting three diseases. An analysis is performed for this test bench on the different pairs of criteria both for the bi- and tricriteria optimization: large portfolios cause resource queues and delays time to launch and are eliminated by the bi- and tricriteria optimization strategy. The optimization strategy is thus interesting to detect the sequence candidates. Time is an important criterion to consider simultaneously with NPV and risk criteria. The order in which drugs are released in the pipeline is of great importance as with scheduling problems

    Safety Aware Vehicle Routing Algorithm, A Weighted Sum Approach

    Get PDF
    Driving is an essential part of work life for many people. Although driving can be enjoyable and pleasant, it can also be stressful and dangerous. Many people around the world are killed or seriously injured while driving. According to the World Health Organization (WHO), about 1.25 million people die each year as a result of road traffic crashes. Road traffic injuries are also the leading cause of death among young people. To prevent traffic injuries, governments must address road safety issues, an endeavor that requires involvement from multiple sectors (transport, police, health, education). Effective intervention should include designing safer infrastructure and incorporating road safety features into land-use and transport planning. The aim of this research is to design an algorithm to help drivers find the safest path between two locations. Such an algorithm can be used to find the safest path for a school bus travelling between bus stops, a heavy truck carrying inflammable materials, poison gas, or explosive cargo, or any driver who wants to avoid roads with higher numbers of accidents. In these applications, a path is safe if the danger factor on either side of the path is no more than a given upper bound. Since travel time is another important consideration for all drivers, the suggested algorithm utilizes traffic data to consider travel time when searching for the safest route. The key achievements of the work presented in this thesis are summarized as follows. Defining the Safest and Quickest Path Problem (SQPP), in which the goal is to find a short and low-risk path between two locations in a road network at a given point of time. Current methods for representing road networks, travel times and safety level were investigated. Two approaches to defining road safety level were identified, and some methods in each approach were presented. An intensive review of traffic routing algorithms was conducted to identify the most well-known algorithms. An empirical study was also conducted to evaluate the performance of some routing algorithms, using metrics such as scalability and computation time. This research approaches the SQPP problem as a bi-objective Shortest Path Problem (SPP), for which the proposed Safety Aware Algorithm (SAA) aims to output one quickest and safest route. The experiments using this algorithm demonstrate its efficacy and practical applicability

    Heuristic Approach to Job Scheduling in a Small Scale Groundnut Oil Processing Firm in Nigeria

    Get PDF
    Groundnut is an important legume cash crop for tropical farmers and its seeds contain high amounts of edible oil (43-55%) and protein (25-28%). This paper developed a framework for the scheduling of activities (jobs) in small scale groundnut oil processing firm in Nigeria. The research problem is addressed using makespan as a measure of performance with CDS, A1 and Usual Serial Order (USO) heuristics solution methods. Findings reveal that A1 and CDS heuristics are preferred to the traditional USO methods. Also, the mean of A1 (27.11) heuristic, followed by CDS (27.22) heuristics, gives the best makespan results while the USO (31.52) gives the worst result. This paper thus presents a framework that could be beneficial to stakeholders in the Groundnut oil processing industry towards improved customer’s satisfaction, less idle time, and profit optimization. Keywords: Groundnut, small enterprises, scheduling of orders, makespans, optimum results

    Journal of Telecommunications and Information Technology, 2010, nr 3

    Get PDF
    kwartalni

    Keyword search in graphs, relational databases and social networks

    Get PDF
    Keyword search, a well known mechanism for retrieving relevant information from a set of documents, has recently been studied for extracting information from structured data (e.g., relational databases and XML documents). It offers an alternative way to query languages (e.g., SQL) to explore databases, which is effective for lay users who may not be familiar with the database schema or the query language. This dissertation addresses some issues in keyword search in structured data. Namely, novel solutions to existing problems in keyword search in graphs or relational databases are proposed. In addition, a problem related to graph keyword search, team formation in social networks, is studied. The dissertation consists of four parts. The first part addresses keyword search over a graph which finds a substructure of the graph containing all or some of the query keywords. Current methods for keyword search over graphs may produce answers in which some content nodes (i.e., nodes that contain input keywords) are not very close to each other. In addition, current methods explore both content and non-content nodes while searching for the result and are thus both time and memory consuming for large graphs. To address the above problems, we propose algorithms for finding r-cliques in graphs. An r-clique is a group of content nodes that cover all the input keywords and the distance between each pair of nodes is less than or equal to r. Two approximation algorithms that produce r-cliques with a bounded approximation ratio in polynomial delay are proposed. In the second part, the problem of duplication-free and minimal keyword search in graphs is studied. Current methods for keyword search in graphs may produce duplicate answers that contain the same set of content nodes. In addition, an answer found by these methods may not be minimal in the sense that some of the nodes in the answer may contain query keywords that are all covered by other nodes in the answer. Removing these nodes does not change the coverage of the answer but can make the answer more compact. We define the problem of finding duplication-free and minimal answers, and propose algorithms for finding such answers efficiently. Meaningful keyword search in relational databases is the subject of the third part of this dissertation. Keyword search over relational databases returns a join tree spanning tuples containing the query keywords. As many answers of varying quality can be found, and the user is often only interested in seeing the·top-k answers, how to gauge the relevance of answers to rank them is of paramount importance. This becomes more pertinent for databases with large and complex schemas. We focus on the relevance of join trees as the fundamental means to rank the answers. We devise means to measure relevance of relations and foreign keys in the schema over the information content of the database. The problem of keyword search over graph data is similar to the problem of team formation in social networks. In this setting, keywords represent skills and the nodes in a graph represent the experts that possess skills. Given an expert network, in which a node represents an expert that has a cost for using the expert service and an edge represents the communication cost between the two corresponding experts, we tackle the problem of finding a team of experts that covers a set of required skills and also minimizes the communication cost as well as the personnel cost of the team. We propose two types of approximation algorithms to solve this bi-criteria problem in the fourth part of this dissertation

    A bi-objective turning restriction design problem in urban road networks

    Get PDF
    postprin

    The bi-objective travelling salesman problem with profits and its connection to computer networks.

    Get PDF
    This is an interdisciplinary work in Computer Science and Operational Research. As it is well known, these two very important research fields are strictly connected. Among other aspects, one of the main areas where this interplay is strongly evident is Networking. As far as most recent decades have seen a constant growing of every kind of network computer connections, the need for advanced algorithms that help in optimizing the network performances became extremely relevant. Classical Optimization-based approaches have been deeply studied and applied since long time. However, the technology evolution asks for more flexible and advanced algorithmic approaches to model increasingly complex network configurations. In this thesis we study an extension of the well known Traveling Salesman Problem (TSP): the Traveling Salesman Problem with Profits (TSPP). In this generalization, a profit is associated with each vertex and it is not necessary to visit all vertices. The goal is to determine a route through a subset of nodes that simultaneously minimizes the travel cost and maximizes the collected profit. The TSPP models the problem of sending a piece of information through a network where, in addition to the sending costs, it is also important to consider what “profit” this information can get during its routing. Because of its formulation, the right way to tackled the TSPP is by Multiobjective Optimization algorithms. Within this context, the aim of this work is to study new ways to solve the problem in both the exact and the approximated settings, giving all feasible instruments that can help to solve it, and to provide experimental insights into feasible networking instances

    A Bicriteria Simulated Annealing Algorithm for Scheduling Jobs on Parallel Machines with Sequence Dependent Setup Times

    Get PDF
    The study considers the scheduling problem of identical parallel machines subject to minimization of the maximum completion time and the maximum tardiness expressed in a linear convex objective function. The maximum completion time or makespan is the date when the last job to be completed leaves the system. The maximum tardiness is indicated by the job that is completed with the longest delay relative its due date. Minimizing both criteria can help assuring a high utilization of the production system as well as a high level of service towards the client. Due to the complexity of the problem, a Simulated Annealing (SA) heuristic has been implemented to be able to obtain an efficient solution in a reasonable running time. A set of n jobs is assigned, to one of the m identical parallel machines. Each job is processed in only one operation before its completion after which it leaves the system. Constraints, such as due dates for each job and setup times for the machines, are considered. The resolution procedure consists of two phases and begins with an initial solution generator. Then a SA heuristic is applied for further improvement of the solution. 4 generators are used to create an initial solution and 3 to generate neighbour solutions. To test and verify the performance of the proposed resolution procedure, a computational experimentation has been realized on a set of test problems generated ad-hoc
    corecore