29,102 research outputs found
PROFET: modeling system performance and energy without simulating the CPU
The approaching end of DRAM scaling and expansion of emerging memory technologies is motivating a lot of research in future memory systems. Novel memory systems are typically explored by hardware simulators that are slow and often have a simplified or obsolete abstraction of the CPU. This study presents PROFET, an analytical model that predicts how an application's performance and energy consumption changes when it is executed on different memory systems. The model is based on instrumentation of an application execution on actual hardware, so it already takes into account CPU microarchitectural details such as the data prefetcher and out-of-order engine. PROFET is evaluated on two real platforms: Sandy Bridge-EP E5-2670 and Knights Landing Xeon Phi platforms with various memory configurations. The evaluation results show that PROFET's predictions are accurate, typically with only 2% difference from the values measured on actual hardware. We release the PROFET source code and all input data required for memory system and application profiling. The released package can be seamlessly installed and used on high-end Intel platforms.Peer ReviewedPostprint (author's final draft
AutoAccel: Automated Accelerator Generation and Optimization with Composable, Parallel and Pipeline Architecture
CPU-FPGA heterogeneous architectures are attracting ever-increasing attention
in an attempt to advance computational capabilities and energy efficiency in
today's datacenters. These architectures provide programmers with the ability
to reprogram the FPGAs for flexible acceleration of many workloads.
Nonetheless, this advantage is often overshadowed by the poor programmability
of FPGAs whose programming is conventionally a RTL design practice. Although
recent advances in high-level synthesis (HLS) significantly improve the FPGA
programmability, it still leaves programmers facing the challenge of
identifying the optimal design configuration in a tremendous design space.
This paper aims to address this challenge and pave the path from software
programs towards high-quality FPGA accelerators. Specifically, we first propose
the composable, parallel and pipeline (CPP) microarchitecture as a template of
accelerator designs. Such a well-defined template is able to support efficient
accelerator designs for a broad class of computation kernels, and more
importantly, drastically reduce the design space. Also, we introduce an
analytical model to capture the performance and resource trade-offs among
different design configurations of the CPP microarchitecture, which lays the
foundation for fast design space exploration. On top of the CPP
microarchitecture and its analytical model, we develop the AutoAccel framework
to make the entire accelerator generation automated. AutoAccel accepts a
software program as an input and performs a series of code transformations
based on the result of the analytical-model-based design space exploration to
construct the desired CPP microarchitecture. Our experiments show that the
AutoAccel-generated accelerators outperform their corresponding software
implementations by an average of 72x for a broad class of computation kernels
Avoiding Aliasing in Allan Variance: an Application to Fiber Link Data Analysis
Optical fiber links are known as the most performing tools to transfer
ultrastable frequency reference signals. However, these signals are affected by
phase noise up to bandwidths of several kilohertz and a careful data processing
strategy is required to properly estimate the uncertainty. This aspect is often
overlooked and a number of approaches have been proposed to implicitly deal
with it. Here, we face this issue in terms of aliasing and show how typical
tools of signal analysis can be adapted to the evaluation of optical fiber
links performance. In this way, it is possible to use the Allan variance as
estimator of stability and there is no need to introduce other estimators. The
general rules we derive can be extended to all optical links. As an example, we
apply this method to the experimental data we obtained on a 1284 km coherent
optical link for frequency dissemination, which we realized in Italy
Control technology overview in CSI
A brief control technology overview is given in Control Structures Interaction (CSI) by illustrating that many future NASA mission present significant challenges as represented by missions having a significantly increased number of important system states which may require control and by identifying key CSI technology needs. The JPL CSI related technology developments are discussed to illustrate that some of the identified control needs are being pursued. Since experimental confirmation of the assumptions inherent in the CSI technology is critically important to establishing its readiness for space program applications, the areas of ground and flight validation require high priority
A multisensing setup for the intelligent tire monitoring
The present paper offers the chance to experimentally measure, for the first time, the internal
tire strain by optical fiber sensors during the tire rolling in real operating conditions. The phenomena
that take place during the tire rolling are in fact far from being completely understood. Despite several
models available in the technical literature, there is not a correspondently large set of experimental
observations. The paper includes the detailed description of the new multi-sensing technology for an
ongoing vehicle measurement, which the research group has developed in the context of the project
OPTYRE. The experimental apparatus is mainly based on the use of optical fibers with embedded
Fiber Bragg Gratings sensors for the acquisition of the circumferential tire strain. Other sensors are
also installed on the tire, such as a phonic wheel, a uniaxial accelerometer, and a dynamic temperature
sensor. The acquired information is used as input variables in dedicated algorithms that allow the
identification of key parameters, such as the dynamic contact patch, instantaneous dissipation and
instantaneous grip. The OPTYRE project brings a contribution into the field of experimental grip
monitoring of wheeled vehicles, with implications both on passive and active safety characteristics of
cars and motorbikes
An Overview on Application of Machine Learning Techniques in Optical Networks
Today's telecommunication networks have become sources of enormous amounts of
widely heterogeneous data. This information can be retrieved from network
traffic traces, network alarms, signal quality indicators, users' behavioral
data, etc. Advanced mathematical tools are required to extract meaningful
information from these data and take decisions pertaining to the proper
functioning of the networks from the network-generated data. Among these
mathematical tools, Machine Learning (ML) is regarded as one of the most
promising methodological approaches to perform network-data analysis and enable
automated network self-configuration and fault management. The adoption of ML
techniques in the field of optical communication networks is motivated by the
unprecedented growth of network complexity faced by optical networks in the
last few years. Such complexity increase is due to the introduction of a huge
number of adjustable and interdependent system parameters (e.g., routing
configurations, modulation format, symbol rate, coding schemes, etc.) that are
enabled by the usage of coherent transmission/reception technologies, advanced
digital signal processing and compensation of nonlinear effects in optical
fiber propagation. In this paper we provide an overview of the application of
ML to optical communications and networking. We classify and survey relevant
literature dealing with the topic, and we also provide an introductory tutorial
on ML for researchers and practitioners interested in this field. Although a
good number of research papers have recently appeared, the application of ML to
optical networks is still in its infancy: to stimulate further work in this
area, we conclude the paper proposing new possible research directions
MR-BART: Multi-Rate Available Bandwidth Estimation in Real-Time
In this paper, we propose Multi-Rate Bandwidth Available in Real Time
(MR-BART) to estimate the end-to-end Available Bandwidth (AB) of a network
path. The proposed scheme is an extension of the Bandwidth Available in Real
Time (BART) which employs multi-rate (MR) probe packet sequences with Kalman
filtering. Comparing to BART, we show that the proposed method is more robust
and converges faster than that of BART and achieves a more AB accurate
estimation. Furthermore, we analyze the estimation error in MR-BART and obtain
analytical formula and empirical expression for the AB estimation error based
on the system parameters.Comment: 12 Pages (Two columns), 14 Figures, 4 Tables
- …