7,500 research outputs found

    XNOR Neural Engine: a Hardware Accelerator IP for 21.6 fJ/op Binary Neural Network Inference

    Full text link
    Binary Neural Networks (BNNs) are promising to deliver accuracy comparable to conventional deep neural networks at a fraction of the cost in terms of memory and energy. In this paper, we introduce the XNOR Neural Engine (XNE), a fully digital configurable hardware accelerator IP for BNNs, integrated within a microcontroller unit (MCU) equipped with an autonomous I/O subsystem and hybrid SRAM / standard cell memory. The XNE is able to fully compute convolutional and dense layers in autonomy or in cooperation with the core in the MCU to realize more complex behaviors. We show post-synthesis results in 65nm and 22nm technology for the XNE IP and post-layout results in 22nm for the full MCU indicating that this system can drop the energy cost per binary operation to 21.6fJ per operation at 0.4V, and at the same time is flexible and performant enough to execute state-of-the-art BNN topologies such as ResNet-34 in less than 2.2mJ per frame at 8.9 fps.Comment: 11 pages, 8 figures, 2 tables, 3 listings. Accepted for presentation at CODES'18 and for publication in IEEE Transactions on Computer-Aided Design of Circuits and Systems (TCAD) as part of the ESWEEK-TCAD special issu

    Energy-efficient resource-provisioning algorithms for optical clouds

    Get PDF
    Rising energy costs and climate change have led to an increased concern for energy efficiency (EE). As information and communication technology is responsible for about 4% of total energy consumption worldwide, it is essential to devise policies aimed at reducing it. In this paper, we propose a routing and scheduling algorithm for a cloud architecture that targets minimal total energy consumption by enabling switching off unused network and/or information technology (IT) resources, exploiting the cloud-specific anycast principle. A detailed energy model for the entire cloud infrastructure comprising a wide-area optical network and IT resources is provided. This model is used to make a single-step decision on which IT end points to use for a given request, including the routing of the network connection toward these end points. Our simulations quantitatively assess the EE algorithm's potential energy savings but also assess the influence this may have on traditional quality-of-service parameters such as service blocking. Furthermore, we compare the one-step scheduling with traditional scheduling and routing schemes, which calculate the resource provisioning in a two-step approach (selecting first the destination IT end point and subsequently using unicast routing toward it). We show that depending on the offered infrastructure load, our proposed one-step calculation considerably lowers the total energy consumption (reduction up to 50%) compared to the traditional iterative scheduling and routing, especially in low-to medium-load scenarios, without any significant increase in the service blocking
    • …
    corecore