591 research outputs found
Cross-Layer Approaches for an Aging-Aware Design of Nanoscale Microprocessors
Thanks to aggressive scaling of transistor dimensions, computers have revolutionized our life. However, the increasing unreliability of devices fabricated in nanoscale technologies emerged as a major threat for the future success of computers. In particular, accelerated transistor aging is of great importance, as it reduces the lifetime of digital systems. This thesis addresses this challenge by proposing new methods to model, analyze and mitigate aging at microarchitecture-level and above
Hybrid Linux System Modeling with Mixed-Level Simulation
Dissertação de mestrado integrado em Engenharia Electrónica Industrial e ComputadoresWe live in a world where the need for computer-based systems with better performances
is growing fast, and part of these systems are embedded systems. This
kind of systems are everywhere around us, and we use them everyday even without
noticing. Nevertheless, there are issues related to embedded systems in what comes
to real-time requirements, because the failure of such systems can be harmful
to the user or its environment.
For this reason, a common technique to meet real-time requirements in difficult
scenarios is accelerating software applications by using parallelization techniques
and dedicated hardware components. This dissertations’ goal is to adopt a methodology
of hardware-software co-design aided by co-simulation, making the design
flow more efficient and reliable. An isolated validation does not guarantee integral
system functionality, but the use of an integrated co-simulation environment
allows detecting system problems before moving to the physical implementation.
In this dissertation, an integrated co-simulation environment will be developed,
using the Quick EMUlator (QEMU) as a tool for emulating embedded software
platforms in a Linux-based environment. A SystemVerilog Direct Programming
Interface (DPI) Library was developed in order to allow SystemVerilog simulators
that support DPI to perform co-simulation with QEMU. A library for DLL
blocks was also developed in order to allow PSIMR to communicate with QEMU.
Together with QEMU, these libraries open up the possibility to co-simulate several
parts of a system that includes power electronics and hardware acceleration
together with an emulated embedded platform.
In order to validate the functionality of the developed co-simulation environment,
a demonstration application scenario was developed following a design flow that
takes advantage of the mentioned simulation environment capabilities.Vivemos num mundo em que a procura por sistemas computer-based com desempenhos
cada vez melhores domina o mercado. Estamos rodeados por este tipo de
sistemas, usando-os todos os dias sem nos apercebermos disso, sendo grande parte
deles sistemas embebidos. Ainda assim, existem problemas relacionados com os
sistemas embebidos no que toca aos requisitos de tempo-real, porque uma falha
destes sistemas pode ser perigosa para o utilizador ou o ambiente que o rodeia.
Devido a isto, uma técnica comum para se conseguir cumprir os requisitos de
tempo-real em aplicações críticas é a aceleração de aplicações de software, utilizando
técnicas de paralelização e o uso de componentes de hardware dedicados.
O objetivo desta dissertação é adotar uma metodologia de co-design de hardwaresoftware
apoiada em co-simulação, tornando o design flow mais eficiente e fiável.
Uma validação isolada não garante a funcionalidade do sistema completo, mas a
utilização de um ambiente de co-simulação permite detetar problemas no sistema
antes deste ser implementado na plataforma alvo.
Nesta dissertação será desenvolvido um ambiente de co-simulação usando o QEMU
como emulador para as plataformas de software "embebido" baseadas em Linux.
Uma biblioteca para SystemVerilog DPI foi desenvolvida, que permite a co-simulação
entre o QEMU e simuladores de Register-Transfer Level (RTL) que suportem SystemVerilog.
Foi também desenvolvida uma biblioteca para os blocos Dynamic Link
Library (DLL) do PSIMR , de modo a permitir a ligação ao QEMU. Em conjunto,
as bibliotecas desenvolvidas permitem a co-simulação de diversas partes do sistema,
nomeadamente do hardware de eletrónica de potência e dos aceleradores de
hardware, juntamente com a plataforma embebida emulada no QEMU.Para validar as funcionalidades do ambiente de co-simulação desenvolvido, foi explorado
um cenário de aplicação que tem por base esse mesmo ambiente
Doubly-fed induction generator used in wind energy
Wound-rotor induction generator has numerous advantages in wind power generation over other generators. One scheme for wound-rotor induction generator is realized when a converter cascade is used between the slip-ring terminals and the utility grid to control the rotor power. This configuration is called the doubly-fed induction generator (DFIG). In this work, a novel induction machine model is developed. This model includes the saturation in the main and leakage flux paths. It shows that the model which considers the saturation effects gives more realistic results. A new technique, which was developed for synchronous machines, was applied to experimentally measure the stator and rotor leakage inductance saturation characteristics on the induction machine.
A vector control scheme is developed to control the rotor side voltage-source converter. Vector control allows decoupled or independent control of both active and reactive power of DFIG. These techniques are based on the theory of controlling the B- and q- axes components of voltage or current in different reference frames. In this work, the stator flux oriented rotor current control, with decoupled control of active and reactive power, is adopted. This scheme allows the independent control of the generated active and reactive power as well as the rotor speed to track the maximum wind power point. Conventionally, the controller type used in vector controllers is of the PI type with a fixed proportional and integral gain. In this work, different intelligent schemes by which the controller can change its behavior are proposed. The first scheme is an adaptive gain scheduler which utilizes different characteristics to generate the variation in the proportional and the integral gains. The second scheme is a fuzzy logic gain scheduler and the third is a neuro-fuzzy controller. The transient responses using the above mentioned schemes are compared analytically and experimentally. It has been found that although the fuzzy logic and neuro-fuzzy schemes are more complicated and have many parameters; this complication provides a higher degree of freedom in tuning the controller which is evident in giving much better system performance. Finally, the simulation results were experimentally verified by building the experimental setup and implementing the developed control schemes
Recommended from our members
On thermal sensor calibration and software techniques for many-core thermal management
The high power density of a many-core processor results in increased temperature which negatively impacts system reliability and performance. Dynamic thermal management applies thermal-aware techniques at run time to avoid overheating using temperature information collected from on-chip thermal sensors. Temperature sensing and thermal control schemes are two critical technologies for successfully maintaining thermal safety. In this dissertation, on-line thermal sensor calibration schemes are developed to provide accurate temperature information.
Software-based dynamic thermal management techniques are proposed using calibrated thermal sensors. Due to process variation and silicon aging, on-chip thermal sensors require periodic calibration before use in DTM. However, the calibration cost for thermal sensors can be prohibitively high as the number of on-chip sensors increases. Linear models which are suitable for on-line calculation are employed to estimate temperatures at multiple sensor locations using performance counters. The estimated temperature and the actual sensor thermal profile show a very high similarity with correlation coefficient ~0.9 for SPLASH2 and SPEC2000 benchmarks.
A calibration approach is proposed to combine potentially inaccurate temperature values obtained from two sources: thermal sensor readings and temperature estimations. A data fusion strategy based on Bayesian inference, which combines information from these two sources, is demonstrated. The result shows the strategy can effectively recalibrate sensor readings in response to inaccuracies caused by process variation and environmental noise. The average absolute error of the corrected sensor temperature readings is
A dynamic task allocation strategy is proposed to address localized overheating in many-core systems. Our approach employs reinforcement learning, a dynamic machine learning algorithm that performs task allocation based on current temperatures and a prediction regarding which assignment will minimize the peak temperature. Our results show that the proposed technique is fast (scheduling performed in \u3c1 \u3ems) and can efficiently reduce peak temperature by up to 8 degree C in a 49-core processor (6% on average) versus a leading competing task allocation approach for a series of SPLASH-2 benchmarks. Reinforcement learning has also been applied to 3D integrated circuits to allocate tasks with thermal awareness
Space station automation of common module power management and distribution
The purpose is to automate a breadboard level Power Management and Distribution (PMAD) system which possesses many functional characteristics of a specified Space Station power system. The automation system was built upon 20 kHz ac source with redundancy of the power buses. There are two power distribution control units which furnish power to six load centers which in turn enable load circuits based upon a system generated schedule. The progress in building this specified autonomous system is described. Automation of Space Station Module PMAD was accomplished by segmenting the complete task in the following four independent tasks: (1) develop a detailed approach for PMAD automation; (2) define the software and hardware elements of automation; (3) develop the automation system for the PMAD breadboard; and (4) select an appropriate host processing environment
Design and Analysis of Dynamic Thermal Management in Chip Multiprocessors (CMPs)
Chip Multiprocessors (CMPs) have been prevailing in the modern microprocessor
market. As the significant heat is converted by the ever-increasing power density and
current leakage, the raised operating temperature in a chip has already threatened
the system?s reliability and led the thermal control to be one of the most important
issues needed to be addressed immediately in chip designs. Due to the cost and
complexity of designing thermal packaging, many Dynamic Thermal Management
(DTM) schemes have been widely adopted in modern processors.
In this study, we focus on developing a simple and accurate thermal model,
which provides a scheduling decision for running tasks. And we show how to design
an efficient DTM scheme with negligible performance overhead. First, we propose an
efficient DTM scheme for multimedia applications that tackles the thermal control
problem in a unified manner. A DTM scheme for multimedia applications makes soft
realtime scheduling decisions based on statistical characteristics of multimedia applications.
Specifically, we model application execution characteristics as the probability
distribution of the number of cycles required to decode frames. Our DTM scheme
for multimedia applications has been implemented on Linux in two mobile processors
providing variable clock frequencies in an Intel Pentium-M processor and an Intel Atom processor. In order to evaluate the performance of the proposed DTM scheme,
we exploit two major codecs, MPEG-4 and H.264/AVC based on various frame resolutions.
Our results show that our DTM scheme for multimedia applications lowers
the overall temperature by 4 degrees C and the peak temperature by 6 degrees C (up to 10 degrees C),
while maintaining frame drop ratio under 5% compared to existing DTM schemes
for multimedia applications. Second, we propose a lightweight online workload estimation
using the cumulative distribution function and architectural information via
Performance Monitoring Counters (PMC) to observe the processes dynamic workload
behaviors. We also present an accurate thermal model for CMP architectures to analyze
the thermal correlation effects by profiling the thermal impacts from neighboring
cores under the specific workload. Hence, according to the estimated workload characteristics
and thermal correlation effects, we can estimate the future temperature of
each core more accurately.
We implement a DTM scheme considering workload characteristics and thermal
correlation effects on real machines, an Intel Quad-Core Q6600 system and Dell
PowerEdge 2950 (dual Intel Xeon E5310 Quad-Core) system, running applications
ranging from multimedia applications to several benchmarks. Experiments results
show that our DTM scheme reduces the peak temperature by 8% with 0.54% performance
overhead compared to Linux Standard Scheduler, while existing DTM schemes
reduce peak temperature by 4% with up to 50% performance overhead
A Processor Extension for Cycle-Accurate Real-Time Software
Certain hard real-time tasks demand precise timing of events, but the usual software solution of periodic interrupts driving a scheduler only provides precision in the millisecond range. NOP-insertion can provide higher precision, but is tedious to do manually, requires predictable instruction timing, and works best with simple algorithms. To achieve high-precision timing in software, we propose instruction-level access to cycle-accurate timers. We add an instruction that waits for a timer to expire then reloads it synchronously. Among other things, this provides a way to exactly specify the period of a loop. To validate our approach, we implemented a simple RISC processor with our extension on an FPGA and programmed it to behave like a video controller and an asynchronous serial receiver. Both applications were much easier to write and debug than their hardware counterparts, which took roughly four times as many lines in VHDL. Simple processors with our extension brings software-style development to a class of applications that were once only possible with hardware
Modeling and Dynamic Management of 3D Multicore Systems with Liquid Cooling
Three-dimensional (3D) circuits reduce communication delay in multicore SoCs, and enable efficient integration of cores, memories, sensors, and RF devices. However, vertical integration of layers exacerbates the reliability and thermal problems, and cooling efficiency becomes a limiting factor. Liquid cooling is a solution to overcome the accelerated thermal problems imposed by multi-layer architectures. In this paper, we first provide a 3D thermal simulation model including liquid cooling, supporting both fixed and variable fluid injection rates. Our model has been integrated in HotSpot to study the impact on multicore SoCs. We design and evaluate several dynamic management policies that complement liquid cooling. Our results for 3D multicore SoCs, which are based on a 3D version of UltraSPARC T1, show that thermal management approaches that combine liquid cooling with proactive task allocation are extremely effective in preventing temperature problems. Our proactive management technique provides an additional 75% average reduction in hot spots in comparison to applying only liquid cooling. Furthermore, for systems capable of varying the coolant flow rate at runtime, our feedback controller increases the improvement to 95% on average
- …