4,351 research outputs found

    ALTERNATE MODELS FOR NATURAL GAS TRANSPORTATION SYSTEM PERFORMANCE OPTIMIZATION

    Get PDF
    The Natural Gas market in the U.S is growing rapidly with evidence that the nation has enough shale reserves to power the country for the next century. To ensure continued economic benefits through the use of this environmentally desired energy source, it becomes important to optimize the transportation network system design. Transportation through pipelines is one of the most common methods used to distribute Natural Gas from source to destination. This transportation system, consisting of pipelines, compressors and other supporting equipment, must be optimized, considering all relevant parameters to minimize cost and increase profit. The research presented here improves on the fuel cost minimization models in literature to incorporate pipeline elevation and safety requirements. A new model is proposed to consider the entire transportation network as a single system and optimize it considering all relevant parameters. The optimization model is setup as a mixed integer nonlinear program. The proposed model is used to optimize the pipeline network for a case study, evaluate the model as well as investigate design capacity and installed capacity of pipeline network

    Iterative restricted space search : a solving approach based on hybridization

    Get PDF
    Face à la complexité qui caractérise les problèmes d'optimisation de grande taille l'exploration complète de l'espace des solutions devient rapidement un objectif inaccessible. En effet, à mesure que la taille des problèmes augmente, des méthodes de solution de plus en plus sophistiquées sont exigées afin d'assurer un certain niveau d 'efficacité. Ceci a amené une grande partie de la communauté scientifique vers le développement d'outils spécifiques pour la résolution de problèmes de grande taille tels que les méthodes hybrides. Cependant, malgré les efforts consentis dans le développement d'approches hybrides, la majorité des travaux se sont concentrés sur l'adaptation de deux ou plusieurs méthodes spécifiques, en compensant les points faibles des unes par les points forts des autres ou bien en les adaptant afin de collaborer ensemble. Au meilleur de notre connaissance, aucun travail à date n'à été effectué pour développer un cadre conceptuel pour la résolution efficace de problèmes d'optimisation de grande taille, qui soit à la fois flexible, basé sur l'échange d'information et indépendant des méthodes qui le composent. L'objectif de cette thèse est d'explorer cette avenue de recherche en proposant un cadre conceptuel pour les méthodes hybrides, intitulé la recherche itérative de l'espace restreint, ±Iterative Restricted Space Search (IRSS)>>, dont, la principale idée est la définition et l'exploration successives de régions restreintes de l'espace de solutions. Ces régions, qui contiennent de bonnes solutions et qui sont assez petites pour être complètement explorées, sont appelées espaces restreints "Restricted Spaces (RS)". Ainsi, l'IRSS est une approche de solution générique, basée sur l'interaction de deux phases algorithmiques ayant des objectifs complémentaires. La première phase consiste à identifier une région restreinte intéressante et la deuxième phase consiste à l'explorer. Le schéma hybride de l'approche de solution permet d'alterner entre les deux phases pour un nombre fixe d'itérations ou jusqu'à l'atteinte d'une certaine limite de temps. Les concepts clés associées au développement de ce cadre conceptuel et leur validation seront introduits et validés graduellement dans cette thèse. Ils sont présentés de manière à permettre au lecteur de comprendre les problèmes que nous avons rencontrés en cours de développement et comment les solutions ont été conçues et implémentées. À cette fin, la thèse a été divisée en quatre parties. La première est consacrée à la synthèse de l'état de l'art dans le domaine de recherche sur les méthodes hybrides. Elle présente les principales approches hybrides développées et leurs applications. Une brève description des approches utilisant le concept de restriction d'espace est aussi présentée dans cette partie. La deuxième partie présente les concepts clés de ce cadre conceptuel. Il s'agit du processus d'identification des régions restreintes et des deux phases de recherche. Ces concepts sont mis en oeuvre dans un schéma hybride heuristique et méthode exacte. L'approche a été appliquée à un problème d'ordonnancement avec deux niveaux de décision, relié au contexte des pâtes et papier: "Pulp Production Scheduling Problem". La troisième partie a permit d'approfondir les concepts développés et ajuster les limitations identifiées dans la deuxième partie, en proposant une recherche itérative appliquée pour l'exploration de RS de grande taille et une structure en arbre binaire pour l'exploration de plusieurs RS. Cette structure a l'avantage d'éviter l'exploration d 'un espace déjà exploré précédemment tout en assurant une diversification naturelle à la méthode. Cette extension de la méthode a été testée sur un problème de localisation et d'allocation en utilisant un schéma d'hybridation heuristique-exact de manière itérative. La quatrième partie généralise les concepts préalablement développés et conçoit un cadre général qui est flexible, indépendant des méthodes utilisées et basé sur un échange d'informations entre les phases. Ce cadre a l'avantage d'être général et pourrait être appliqué à une large gamme de problèmes

    An efficient design space exploration framework to optimize power-efficient heterogeneous many-core multi-threading embedded processor architectures

    Get PDF
    By the middle of this decade, uniprocessor architecture performance had hit a roadblock due to a combination of factors, such as excessive power dissipation due to high operating frequencies, growing memory access latencies, diminishing returns on deeper instruction pipelines, and a saturation of available instruction level parallelism in applications. An attractive and viable alternative embraced by all the processor vendors was multi-core architectures where throughput is improved by using micro-architectural features such as multiple processor cores, interconnects and low latency shared caches integrated on a single chip. The individual cores are often simpler than uniprocessor counterparts, use hardware multi-threading to exploit thread-level parallelism and latency hiding and typically achieve better performance-power figures. The overwhelming success of the multi-core microprocessors in both high performance and embedded computing platforms motivated chip architects to dramatically scale the multi-core processors to many-cores which will include hundreds of cores on-chip to further improve throughput. With such complex large scale architectures however, several key design issues need to be addressed. First, a wide range of micro- architectural parameters such as L1 caches, load/store queues, shared cache structures and interconnection topologies and non-linear interactions between them define a vast non-linear multi-variate micro-architectural design space of many-core processors; the traditional method of using extensive in-loop simulation to explore the design space is simply not practical. Second, to accurately evaluate the performance (measured in terms of cycles per instruction (CPI)) of a candidate design, the contention at the shared cache must be accounted in addition to cycle-by-cycle behavior of the large number of cores which superlinearly increases the number of simulation cycles per iteration of the design exploration. Third, single thread performance does not scale linearly with number of hardware threads per core and number of cores due to memory wall effect. This means that at every step of the design process designers must ensure that single thread performance is not unacceptably slowed down while increasing overall throughput. While all these factors affect design decisions in both high performance and embedded many-core processors, the design of embedded processors required for complex embedded applications such as networking, smart power grids, battlefield decision-making, consumer electronics and biomedical devices to name a few, is fundamentally different from its high performance counterpart because of the need to consider (i) low power and (ii) real-time operations. This implies the design objective for embedded many-core processors cannot be to simply maximize performance, but improve it in such a way that overall power dissipation is minimized and all real-time constraints are met. This necessitates additional power estimation models right at the design stage to accurately measure the cost and reliability of all the candidate designs during the exploration phase. In this dissertation, a statistical machine learning (SML) based design exploration framework is presented which employs an execution-driven cycle- accurate simulator to accurately measure power and performance of embedded many-core processors. The embedded many-core processor domain is Network Processors (NePs) used to processed network IP packets. Future generation NePs required to operate at terabits per second network speeds captures all the aspects of a complex embedded application consisting of shared data structures, large volume of compute-intensive and data-intensive real-time bound tasks and a high level of task (packet) level parallelism. Statistical machine learning (SML) is used to efficiently model performance and power of candidate designs in terms of wide ranges of micro-architectural parameters. The method inherently minimizes number of in-loop simulations in the exploration framework and also efficiently captures the non-linear interactions between the micro-architectural design parameters. To ensure scalability, the design space is partitioned into (i) core-level micro-architectural parameters to optimize single core architectures subject to the real-time constraints and (ii) shared memory level micro- architectural parameters to explore the shared interconnection network and shared cache memory architectures and achieves overall optimality. The cost function of our exploration algorithm is the total power dissipation which is minimized, subject to the constraints of real-time throughput (as determined from the terabit optical network router line-speed) required in IP packet processing embedded application

    Lost in optimisation of water distribution systems? A literature review of system operation

    Get PDF
    This is the author accepted manuscript. The final version is available from Elsevier via the DOI in this record.Optimisation of the operation of water distribution systems has been an active research field for almost half a century. It has focused mainly on optimal pump operation to minimise pumping costs and optimal water quality management to ensure that standards at customer nodes are met. This paper provides a systematic review by bringing together over two hundred publications from the past three decades, which are relevant to operational optimisation of water distribution systems, particularly optimal pump operation, valve control and system operation for water quality purposes of both urban drinking and regional multiquality water distribution systems. Uniquely, it also contains substantial and thorough information for over one hundred publications in a tabular form, which lists optimisation models inclusive of objectives, constraints, decision variables, solution methodologies used and other details. Research challenges in terms of simulation models, optimisation model formulation, selection of optimisation method and postprocessing needs have also been identified

    Training and Serving System of Foundation Models: A Comprehensive Survey

    Full text link
    Foundation models (e.g., ChatGPT, DALL-E, PengCheng Mind, PanGu-ÎŁ\Sigma) have demonstrated extraordinary performance in key technological areas, such as natural language processing and visual recognition, and have become the mainstream trend of artificial general intelligence. This has led more and more major technology giants to dedicate significant human and financial resources to actively develop their foundation model systems, which drives continuous growth of these models' parameters. As a result, the training and serving of these models have posed significant challenges, including substantial computing power, memory consumption, bandwidth demands, etc. Therefore, employing efficient training and serving strategies becomes particularly crucial. Many researchers have actively explored and proposed effective methods. So, a comprehensive survey of them is essential for system developers and researchers. This paper extensively explores the methods employed in training and serving foundation models from various perspectives. It provides a detailed categorization of these state-of-the-art methods, including finer aspects such as network, computing, and storage. Additionally, the paper summarizes the challenges and presents a perspective on the future development direction of foundation model systems. Through comprehensive discussion and analysis, it hopes to provide a solid theoretical basis and practical guidance for future research and applications, promoting continuous innovation and development in foundation model systems

    Detection and Localization of Leaks in Water Networks

    Get PDF
    Today, 844 million humans around the world have no access to safe drinking water. Furthermore, every 90 seconds, one child dies from water-related illnesses. Major cities lose 15% - 50% of their water and, in some cases, losses may reach up to 70%, mostly due to leaks. Therefore, it is paramount to preserve water as an invaluable resource through water networks, particularly in large cities in which leak repair may cause disruption. Municipalities usually tackle leak problems using various detection systems and technologies, often long after leaks occur; however, such efforts are not enough to detect leaks at early stages. Therefore, the main objectives of the present research are to develop and validate a leak detection system and to optimize leak repair prioritization. The development of the leak detection models goes through several phases: (1) technology and device selection, (2) experimental work, (3) signal analysis, (4) selection of parameters, (5) machine learning model development and (6) validation of developed models. To detect leaks, vibration signals are collected through a variety of controlled experiments on PVC and ductile iron pipelines using wireless accelerometers, i.e., micro-electronic mechanical sensors (MEMS). The signals are analyzed to pinpoint leaks in water pipelines. Similarly, acoustic signals are collected from a pilot project in the city of Montreal, using noise loggers as another detection technology. The collected signals are also analyzed to detect and pinpoint the leaks. The leak detection system has presented promising results using both technologies. The developed MEMS model is capable of accurately pinpointing leaks within 12 centimeters from the exact location. Comparatively, for noise loggers, the developed model can detect the exact leak location within a 25-cm radius for an actual leak. The leak repair prioritization model uses two optimization techniques: (1) a well-known genetic algorithm and (2) a newly innovative Lazy Serpent Algorithm that is developed in the present research. The Lazy Serpent Algorithm has proved capable of surpassing the genetic algorithm in determining a more optimal schedule using much less computation time. The developed research proves that automated real-time leak detection is possible and can help governments save water resource and funds. The developed research proves the viability of accelerometers as a standalone leak detection technology and opens the door for further research and experimentations. The leak detection system model helps municipalities and water resource agencies rapidly detect leaks when they occur in real-time. The developed pinpointing models facilitate the leak repair process by precisely determine the leak location where the repair works should be conducted. The Lazy Serpent Algorithm helps municipalities better distribute their resources to maximize their desired benefits

    A Probabilistic Approach for the System-Level Design of Multi-ASIP Platforms

    Get PDF
    • …
    corecore