7,500 research outputs found
XNOR Neural Engine: a Hardware Accelerator IP for 21.6 fJ/op Binary Neural Network Inference
Binary Neural Networks (BNNs) are promising to deliver accuracy comparable to
conventional deep neural networks at a fraction of the cost in terms of memory
and energy. In this paper, we introduce the XNOR Neural Engine (XNE), a fully
digital configurable hardware accelerator IP for BNNs, integrated within a
microcontroller unit (MCU) equipped with an autonomous I/O subsystem and hybrid
SRAM / standard cell memory. The XNE is able to fully compute convolutional and
dense layers in autonomy or in cooperation with the core in the MCU to realize
more complex behaviors. We show post-synthesis results in 65nm and 22nm
technology for the XNE IP and post-layout results in 22nm for the full MCU
indicating that this system can drop the energy cost per binary operation to
21.6fJ per operation at 0.4V, and at the same time is flexible and performant
enough to execute state-of-the-art BNN topologies such as ResNet-34 in less
than 2.2mJ per frame at 8.9 fps.Comment: 11 pages, 8 figures, 2 tables, 3 listings. Accepted for presentation
at CODES'18 and for publication in IEEE Transactions on Computer-Aided Design
of Circuits and Systems (TCAD) as part of the ESWEEK-TCAD special issu
Energy-efficient resource-provisioning algorithms for optical clouds
Rising energy costs and climate change have led to an increased concern for energy efficiency (EE). As information and communication technology is responsible for about 4% of total energy consumption worldwide, it is essential to devise policies aimed at reducing it. In this paper, we propose a routing and scheduling algorithm for a cloud architecture that targets minimal total energy consumption by enabling switching off unused network and/or information technology (IT) resources, exploiting the cloud-specific anycast principle. A detailed energy model for the entire cloud infrastructure comprising a wide-area optical network and IT resources is provided. This model is used to make a single-step decision on which IT end points to use for a given request, including the routing of the network connection toward these end points. Our simulations quantitatively assess the EE algorithm's potential energy savings but also assess the influence this may have on traditional quality-of-service parameters such as service blocking. Furthermore, we compare the one-step scheduling with traditional scheduling and routing schemes, which calculate the resource provisioning in a two-step approach (selecting first the destination IT end point and subsequently using unicast routing toward it). We show that depending on the offered infrastructure load, our proposed one-step calculation considerably lowers the total energy consumption (reduction up to 50%) compared to the traditional iterative scheduling and routing, especially in low-to medium-load scenarios, without any significant increase in the service blocking
- …