559 research outputs found

    Stochastic Learning Feedback Hybrid Automata for Power Management in Embedded Systems

    Full text link
    In this paper we show that stochastic learning automata based feedback control switching strategy can be used for dynamic power management (DPM) employed at the system level. DPM strategies are usually incorporated at the operating systems of embedded devices to exploit multiple power states available in today\u27s ACPI compliant devices. The idea is to switch between power states depending on the device usage, and since device usage times are not deterministic, probabilistic techniques are often used to create stochastic strategies, or strategies that make decisions based on probabilities of device usage spans. Previous work (Irani et al., 2001) has shown how to approximate the probability distribution of device idle times and dynamically update them, and then use such knowledge in controlling power states. Here, we use stochastic learning automata (SLA) which interacts with the environment to update such probabilities, and then apply techniques similar to (Irani et al., 2001) to optimize power usage with minimal effect on response time for the devices

    Architectures for Adaptive Low-Power Embedded Multimedia Systems

    Get PDF
    This Ph.D. thesis describes novel hardware/software architectures for adaptive low-power embedded multimedia systems. Novel techniques for run-time adaptive energy management are proposed, such that both HW & SW adapt together to react to the unpredictable scenarios. A complete power-aware H.264 video encoder was developed. Comparison with state-of-the-art demonstrates significant energy savings while meeting the performance constraint and keeping the video quality degradation unnoticeable

    Software-controlled processor speed setting for low-power streaming multimedia

    Full text link

    A Hierarchical Framework of Cloud Resource Allocation and Power Management Using Deep Reinforcement Learning

    Full text link
    Automatic decision-making approaches, such as reinforcement learning (RL), have been applied to (partially) solve the resource allocation problem adaptively in the cloud computing system. However, a complete cloud resource allocation framework exhibits high dimensions in state and action spaces, which prohibit the usefulness of traditional RL techniques. In addition, high power consumption has become one of the critical concerns in design and control of cloud computing systems, which degrades system reliability and increases cooling cost. An effective dynamic power management (DPM) policy should minimize power consumption while maintaining performance degradation within an acceptable level. Thus, a joint virtual machine (VM) resource allocation and power management framework is critical to the overall cloud computing system. Moreover, novel solution framework is necessary to address the even higher dimensions in state and action spaces. In this paper, we propose a novel hierarchical framework for solving the overall resource allocation and power management problem in cloud computing systems. The proposed hierarchical framework comprises a global tier for VM resource allocation to the servers and a local tier for distributed power management of local servers. The emerging deep reinforcement learning (DRL) technique, which can deal with complicated control problems with large state space, is adopted to solve the global tier problem. Furthermore, an autoencoder and a novel weight sharing structure are adopted to handle the high-dimensional state space and accelerate the convergence speed. On the other hand, the local tier of distributed server power managements comprises an LSTM based workload predictor and a model-free RL based power manager, operating in a distributed manner.Comment: accepted by 37th IEEE International Conference on Distributed Computing (ICDCS 2017

    Stochastic Learning Feedback Hybrid Automata for Dynamic Power Management in Embedded Systems

    Full text link
    Dynamic power management (DPM) refers to the strategies employed at system level to reduce energy expenditure (i.e. to prolong battery life) in embedded systems. The trade-off involved in DPM techniques is between the reductions of energy consumption and latency suffered by the tasks. Such trade-offs need to be decided at runtime, making DPM an on-line problem. We formulate DPM as a hybrid automaton control problem and integrate stochastic control. The control strategy is learnt dynamically using stochastic learning hybrid automata (SLHA) with feedback learning algorithms. Simulation-based experiments show the expediency of the feedback systems in stationary environments. Further experiments reveal that SLHA attains better trade-offs than several former predictive algorithms under certain trace data

    Power Analysis and Optimization Techniques for Energy Efficient Computer Systems

    Get PDF
    Reducing power consumption has become a major challenge in the design and operation of to-day’s computer systems. This chapter describes different techniques addressing this challenge at different levels of system hardware, such as CPU, memory, and internal interconnection network, as well as at different levels of software components, such as compiler, operating system and user applications. These techniques can be broadly categorized into two types: Design time power analysis versus run-time dynamic power management. Mechanisms in the first category use ana-lytical energy models that are integrated into existing simulators to measure the system’s power consumption and thus help engineers to test power-conscious hardware and software during de-sign time. On the other hand, dynamic power management techniques are applied during run-time, and are used to monitor system workload and adapt the system’s behavior dynamically to save energy

    Variable-rate data sampling for low-power microsystems using modified Adams methods

    Get PDF
    A method for variable-rate data sampling is proposed for the purpose of low-power data acquisition in a small footprint microsystem. The procedure enables energy saving by utilizing dynamic power management techniques and is based on the Adams-Bashforth and Adams-Moulton multistep predictor-corrector methods for ordinary differential equations. Newton-Gregory backward difference interpolation formulae and past value substitution are used to facilitate sample rate changes. It is necessary to store only 2m+1 equispaced past values of t and the corresponding values of y, where y=g(t), and m is the number of steps in the Adams methods. For the purposes of demonstrating the technique, fourth-order methods are used, but it is possible to use higher orders to improve accuracy if required

    A Power-Efficient Methodology for Mapping Applications on Multi-Processor System-on-Chip Architectures

    Get PDF
    This work introduces an application mapping methodology and case study for multi-processor on-chip architectures. Starting from the description of an application in standard sequential code (e.g. in C), first the application is profiled, parallelized when possible, then its components are moved to hardware implementation when necessary to satisfy performance and power constraints. After mapping, with the use of hardware objects to handle concurrency, the application power consumption can be further optimized by a task-based scheduler for the remaining software part, without the need for operating system support. The key contributions of this work are: a methodology for high-level hardware/software partitioning that allows the designer to use the same code for both hardware and software models for simulation, providing nevertheless preliminary estimations for timing and power consumption; and a task-based scheduling algorithm that does not require operating system support. The methodology has been applied to the co-exploration of an industrial case study: an MPEG4 VGA real-time encoder

    Low power JPEG2000 5/3 discrete wavelet transform algorithm and architecture

    Get PDF
    corecore