253 research outputs found

    Diseño de circuitos analógicos y de señal mixta con consideraciones de diseño físico y variabilidad

    Get PDF
    Advances in microelectronic technology has been based on an increasing capacity to integrate transistors, moving this industry to the nanoelectronics realm in recent years. Moore’s Law [1] has predicted (and somehow governed) the growth of the capacity to integrate transistors in a single IC. Nevertheless, while this capacity has grown steadily, the increasing number of design tasks that are involved in the creation of the integrated circuit and their complexity has led to a phenomenon known as the ``design gap´´. This is the difference between what can theoretically be integrated and what can practically be designed. Since the early 2000s, the International Technology Roadmap of Semiconductors (ITRS) reports, published by the Semiconductor Industry Association (SIA), alert about the necessity to limit the growth of the design cost by increasing the productivity of the designer to continue the semiconductor industry’s growth. Design automation arises as a key element to close this ”design gap”. In this sense, electronic design automation (EDA) tools have reached a level of maturity for digital circuits that is far behind the EDA tools that are made for analog circuit design automation. While digital circuits rely, in general, on two stable operation states (which brings inherent robustness against numerous imperfections and interferences, leading to few design constraints like area, speed or power consumption), analog signal processing, on the other hand, demands compliance with lots of constraints (e.g., matching, noise, robustness, ...). The triumph of digital CMOS circuits, thanks to their mentioned robustness, has, ultimately, facilitated the way that circuits can be processed by algorithms, abstraction levels and description languages, as well as how the design information traverse the hierarchical levels of a digital system. The field of analog design automation faces many more difficulties due to the many sources of perturbation, such as the well-know process variability, and the difficulty in treating these systematically, like digital tools can do. In this Thesis, different design flows are proposed, focusing on new design methodologies for analog circuits, thus, trying to close the ”gap” between digital and analog EDA tools. In this chapter, the most important sources for perturbations and their impact on the analog design process are discussed in Section 1.2. The traditional analog design flow is discussed in 1.3. Emerging design methodologies that try to reduce the ”design gap” are presented in Section 1.4 where the key concept of Pareto-Optimal Front (POF) is explained. This concept, brought from the field of economics, models the analog circuit performances into a set of solutions that show the optimal trade-offs among conflicting circuit performances (e.g. DC-gain and unity-gain frequency). Finally, the goals of this thesis are presented in Section 1.5

    Fiabilisation de convertisseurs analogique-numérique à modulation Sigma-Delta

    Get PDF
    This thesis concentrates on reliability-aware methodology development, reliability analysis based on simulation as well as failure prediction of CMOS 65nm analog and mixed signal (AMS) ICs. Sigma-Delta modulators are concerned as the object of reliability study at system level. A hierarchical statistical approach for reliability is proposed to analysis the performance of Sigma-Delta modulators under ageing effects and process variations. Statistical methods are combined into this analysis flow.Ce travail de thèse a porté sur des problèmes de fiabilité de circuits intégrés en technologie CMOS 65 nm, en particulier sur la conception en vue de la fiabilité, la simulation et l'amélioration de la fiabilité. Les mécanismes dominants de vieillissement HCI et NBTI ainsi que la variation du processus ont été étudiés et évalués quantitativement au niveau du circuit et au niveau du système. Ces méthodes ont été appliquées aux modulateurs Sigma-Delta afin de déterminer la fiabilité de ce type de composant qui est très utilisé

    Fiabilisation de Convertisseurs Analogique-Num´erique a Modulation Sigma-Delta

    Get PDF
    Due to the continuously scaling down of CMOS technology, system-on-chips (SoCs) reliability becomes important in sub-90 nm CMOS node. Integrated circuits and systems applied to aerospace, avionic, vehicle transport and biomedicine are highly sensitive to reliability problems such as ageing mechanisms and parametric process variations. Novel SoCs with new materials and architectures of high complexity further aggravate reliability as a critical aspect of process integration. For instance, random and systematic defects as well as parametric process variations have a large influence on quality and yield of the manufactured ICs, right after production. During ICs usage time, time-dependent ageing mechanisms such as negative bias temperature instability (NBTI) and hot carrier injection (HCI) can significantly degrade ICs performance.La fiabilit´e des ICs est d´efinie ainsi : la capacit´e d’un circuit ou un syst`eme int´egr´e `amaintenir ses param`etres durant une p´eriode donn´ee sous des conditions d´efinies. Les rapportsITRS 2011 consid`ere la fiabilit´e comme un aspect critique du processus d’int´egration.Par cons´equent, il faut faire appel des m´ethodologies innovatrices prenant en comptela fiabilit´e afin d’assurer la fonctionnalit´e du SoCs et la fiabilit´e dans les technologiesCMOS `a l’´echelle nanom´etrique. Cela nous permettra de d´evelopper des m´ethodologiesind´ependantes du design et de la technologie CMOS, en revanche, sp´ecialis´ees en fiabilit´e

    An integrated operation and maintenance framework for offshore renewable energy

    Get PDF
    Offshore renewable devices hold a large potential as renewable energy sources, but their deployment costs are still too high compared to those of other technologies. Operation and maintenance, as well as management of the assets, are main contributors to the overall costs of the projects, and decision-support tools in this area are required to decrease the final cost of energy.\\ In this thesis a complete characterisation and optimisation framework for the operation, maintenance and assets management of an offshore renewable farm is presented. The methodology uses known approaches, based on Monte Carlo simulation for the characterisation of the key performance indicators of the offshore renewable farm, and genetic algorithms as a search heuristic for the proposal of improved strategies. These methods, coupled in an integrated framework, constitute a novel and valuable tool to support the decision-making process in this area. The methods developed consider multiple aspects for the accurate description of the problem, including considerations on the reliability of the devices and limitations on the offshore operations dictated by the properties of the maintenance assets. Mechanisms and constraints that influence the maintenance procedures are considered and used to determine the optimal strategy. The models are flexible over a range of offshore renewable technologies, and adaptable to different offshore farm sizes and layouts, as well as maintenance assets and configurations of the devices. The approaches presented demonstrate the potential for cost reduction in the operation and maintenance strategy selection, and highlight the importance of computational tools to improve the profitability of a project while ensuring that satisfactory levels of availability and reliability are preserved. Three case studies to show the benefits of application of such methodologies, as well as the validity of their implementation, are provided. Areas for further development are identified, and suggestions to improve the effectiveness of decision-making tools for the assets management of offshore renewable technologies are provided.European CommissionMojo Ocean Dynamics Ltd. T/A Mojo Maritime Lt

    Datacenter management for on-site intermittent and uncertain renewable energy sources

    Get PDF
    Les technologies de l'information et de la communication sont devenues, au cours des dernières années, un pôle majeur de consommation énergétique avec les conséquences environnementales associées. Dans le même temps, l'émergence du Cloud computing et des grandes plateformes en ligne a causé une augmentation en taille et en nombre des centres de données. Pour réduire leur impact écologique, alimenter ces centres avec des sources d'énergies renouvelables (EnR) apparaît comme une piste de solution. Cependant, certaines EnR telles que les énergies solaires et éoliennes sont liées aux conditions météorologiques, et sont par conséquent intermittentes et incertaines. L'utilisation de batteries ou d'autres dispositifs de stockage est souvent envisagée pour compenser ces variabilités de production. De par leur coût important, économique comme écologique, ainsi que les pertes énergétiques engendrées, l'utilisation de ces dispositifs sans intégration supplémentaire est insuffisante. La consommation électrique d'un centre de données dépend principalement de l'utilisation des ressources de calcul et de communication, qui est déterminée par la charge de travail et les algorithmes d'ordonnancement utilisés. Pour utiliser les EnR efficacement tout en préservant la qualité de service du centre, une gestion coordonnée des ressources informatiques, des sources électriques et du stockage est nécessaire. Il existe une grande diversité de centres de données, ayant différents types de matériel, de charge de travail et d'utilisation. De la même manière, suivant les EnR, les technologies de stockage et les objectifs en termes économiques ou environnementaux, chaque infrastructure électrique est modélisée et gérée différemment des autres. Des travaux existants proposent des méthodes de gestion d'EnR pour des couples bien spécifiques de modèles électriques et informatiques. Cependant, les multiples combinaisons de ces deux parties rendent difficile l'extrapolation de ces approches et de leurs résultats à des infrastructures différentes. Cette thèse explore de nouvelles méthodes pour résoudre ce problème de coordination. Une première contribution reprend un problème d'ordonnancement de tâches en introduisant une abstraction des sources électriques. Un algorithme d'ordonnancement est proposé, prenant les préférences des sources en compte, tout en étant conçu pour être indépendant de leur nature et des objectifs de l'infrastructure électrique. Une seconde contribution étudie le problème de planification de l'énergie d'une manière totalement agnostique des infrastructures considérées. Les ressources informatiques et la gestion de la charge de travail sont encapsulées dans une boîte noire implémentant un ordonnancement sous contrainte de puissance. La même chose s'applique pour le système de gestion des EnR et du stockage, qui agit comme un algorithme d'optimisation d'engagement de sources pour répondre à une demande. Une optimisation coopérative et multiobjectif, basée sur un algorithme évolutionnaire, utilise ces deux boîtes noires afin de trouver les meilleurs compromis entre les objectifs électriques et informatiques. Enfin, une troisième contribution vise les incertitudes de production des EnR pour une infrastructure plus spécifique. En utilisant une formulation en processus de décision markovien (MDP), la structure du problème de décision sous-jacent est étudiée. Pour plusieurs variantes du problème, des méthodes sont proposées afin de trouver les politiques optimales ou des approximations de celles-ci avec une complexité raisonnable.In recent years, information and communication technologies (ICT) became a major energy consumer, with the associated harmful ecological consequences. Indeed, the emergence of Cloud computing and massive Internet companies increased the importance and number of datacenters around the world. In order to mitigate economical and ecological cost, powering datacenters with renewable energy sources (RES) began to appear as a sustainable solution. Some of the commonly used RES, such as solar and wind energies, directly depends on weather conditions. Hence they are both intermittent and partly uncertain. Batteries or other energy storage devices (ESD) are often considered to relieve these issues, but they result in additional energy losses and are too costly to be used alone without more integration. The power consumption of a datacenter is closely tied to the computing resource usage, which in turn depends on its workload and on the algorithms that schedule it. To use RES as efficiently as possible while preserving the quality of service of a datacenter, a coordinated management of computing resources, electrical sources and storage is required. A wide variety of datacenters exists, each with different hardware, workload and purpose. Similarly, each electrical infrastructure is modeled and managed uniquely, depending on the kind of RES used, ESD technologies and operating objectives (cost or environmental impact). Some existing works successfully address this problem by considering a specific couple of electrical and computing models. However, because of this combined diversity, the existing approaches cannot be extrapolated to other infrastructures. This thesis explores novel ways to deal with this coordination problem. A first contribution revisits batch tasks scheduling problem by introducing an abstraction of the power sources. A scheduling algorithm is proposed, taking preferences of electrical sources into account, though designed to be independent from the type of sources and from the goal of the electrical infrastructure (cost, environmental impact, or a mix of both). A second contribution addresses the joint power planning coordination problem in a totally infrastructure-agnostic way. The datacenter computing resources and workload management is considered as a black-box implementing a scheduling under variable power constraint algorithm. The same goes for the electrical sources and storage management system, which acts as a source commitment optimization algorithm. A cooperative multiobjective power planning optimization, based on a multi-objective evolutionary algorithm (MOEA), dialogues with the two black-boxes to find the best trade-offs between electrical and computing internal objectives. Finally, a third contribution focuses on RES production uncertainties in a more specific infrastructure. Based on a Markov Decision Process (MDP) formulation, the structure of the underlying decision problem is studied. For several variants of the problem, tractable methods are proposed to find optimal policies or a bounded approximation

    Automated Optimization Strategies for Horizontal Wellbore and Hydraulic Fracture Stages Placement in Unconventional Gas Reseroirs

    Get PDF
    In the last decades rapid advances in horizontal drilling and hydraulic fracturing technologies ensure production of commercial quantities of natural gas from many unconventional reservoirs. Reservoir management and development strategies for shale and tight gas plays have evolved from ad hoc approaches to more rigorous strategies that involve numerical optimization in presence of multiple economic and production objectives and constraints. Application of an automated integrated optimization framework for placement of horizontal wellbores and transverse hydraulic fracture stages along them has potential of increasing shale gas reserves and projects’ revenue even further. This dissertation introduces a novel integrated evolutionary-based optimization framework for placement of horizontal wellbores and hydraulic fracture stages that allows enhancing production from shale gas formations and provides a solid foundation for future field-scale application once better understanding of shale petrophysics and geomechanics is developed. The proposed optimization workflow is developed and tested in stages. First, we summarize what has been done in the subject field previously by scholars and identify what is missing. Second, we present assumptions for the shale gas simulation model that make our framework and the simulation model applicable. Third, we pre-screen several economic and petrophysical parameters in order to identify the most significant for the subsequent sensitivities analysis. Forth, we develop evolutionary-based optimization strategy for placement of hydraulic fracture stages along a single horizontal wellbore. We investigate how sensitive the optimization results to changes in the key parameters pre-selected during pre-screening. Fifth, we enhance the framework to handle multiple horizontal producers, discuss the conditions when such approach is applicable, and extensively test this integrated workflow on a suite of simulation runs. Finally, we implement and apply multi-objective optimization approach (the improved non-dominated sorting genetic algorithm) to the problem of optimal HF stage placement in shale gas reservoirs and analyze the efficiency of our evolutionary-based optimization scheme in presence of multiple conflicting or non-conflicting objectives. Based on our extensive testing and rigorous formulation of the optimization problem, we find that the chosen evolutionary framework is effective in calculating the optimal number of horizontal wells, the number of HF stages, their specific locations along the wells as well as their half-length. We also conclude that further computational efficiency can be achieved if minimum stage spacing and same chromosome elimination procedure are used. The multi-objective approach has been tested on conflicting and non-conflicting objectives and proved to compute the Pareto optimal front of solutions (or production scenarios) in computationally efficient manner

    MULTI-OBJECTIVE DESIGN AUTOMATION FOR RECONFIGURABLE MULTI-PROCESSOR SYSTEMS

    Get PDF
    Ph.DDOCTOR OF PHILOSOPH

    Natural computing for vehicular networks

    Get PDF
    La presente tesis aborda el diseño inteligente de soluciones para el despliegue de redes vehiculares ad-hoc (vehicular ad hoc networks, VANETs). Estas son redes de comunicación inalámbrica formada principalmente por vehículos y elementos de infraestructura vial. Las VANETs ofrecen la oportunidad para desarrollar aplicaciones revolucionarias en el ámbito de la seguridad y eficiencia vial. Al ser un dominio tan novedoso, existe una serie de cuestiones abiertas, como el diseño de la infraestructura de estaciones base necesaria y el encaminamiento (routing) y difusión (broadcasting) de paquetes de datos, que todavía no han podido resolverse empleando estrategias clásicas. Es por tanto necesario crear y estudiar nuevas técnicas que permitan de forma eficiente, eficaz, robusta y flexible resolver dichos problemas. Este trabajo de tesis doctoral propone el uso de computación inspirada en la naturaleza o Computación Natural (CN) para tratar algunos de los problemas más importantes en el ámbito de las VANETs, porque representan una serie de algoritmos versátiles, flexibles y eficientes para resolver problemas complejos. Además de resolver los problemas VANET en los que nos enfocamos, se han realizado avances en el uso de estas técnicas para que traten estos problemas de forma más eficiente y eficaz. Por último, se han llevado a cabo pruebas reales de concepto empleando vehículos y dispositivos de comunicación reales en la ciudad de Málaga (España). La tesis se ha estructurado en cuatro grandes fases. En la primera fase, se han estudiado los principales fundamentos en los que se basa esta tesis. Para ello se hizo un estudio exhaustivo sobre las tecnologías que emplean las redes vehiculares, para así, identificar sus principales debilidades. A su vez, se ha profundizado en el análisis de la CN como herramienta eficiente para resolver problemas de optimización complejos, y de cómo utilizarla en la resolución de los problemas en VANETs. En la segunda fase, se han abordado cuatro problemas de optimización en redes vehiculares: la transferencia de archivos, el encaminamiento (routing) de paquetes, la difusión (broadcasting) de mensajes y el diseño de la infraestructura de estaciones base necesaria para desplegar redes vehiculares. Para la resolución de dichos problemas se han propuesto diferentes algoritmos CN que se clasifican en algoritmos evolutivos (evolutionary algorithms, EAs), métodos de inteligencia de enjambre (swarm intelligence, SI) y enfriamiento simulado (simulated annealing, SA). Los resultados obtenidos han proporcionado protocolos de han mejorado de forma significativa las comunicaciones en VANETs. En la tercera y última fase, se han realizado experimentos empleando vehículos reales circulando por las carreteras de Málaga y que se comunicaban entre sí. El principal objetivo de estas pruebas ha sido el validar las mejoras que presentan los protocolos que se han optimizado empleando CN. Los resultados obtenidos de las fases segunda y tercera confirman la hipótesis de trabajo, que la CN es una herramienta eficiente para tratar el diseño inteligente en redes vehiculares
    corecore