3,114 research outputs found
Multi-Level Pre-Correlation RFI Flagging for Real-Time Implementation on UniBoard
Because of the denser active use of the spectrum, and because of radio
telescopes higher sensitivity, radio frequency interference (RFI) mitigation
has become a sensitive topic for current and future radio telescope designs.
Even if quite sophisticated approaches have been proposed in the recent years,
the majority of RFI mitigation operational procedures are based on
post-correlation corrupted data flagging. Moreover, given the huge amount of
data delivered by current and next generation radio telescopes, all these RFI
detection procedures have to be at least automatic and, if possible, real-time.
In this paper, the implementation of a real-time pre-correlation RFI
detection and flagging procedure into generic high-performance computing
platforms based on Field Programmable Gate Arrays (FPGA) is described,
simulated and tested. One of these boards, UniBoard, developed under a Joint
Research Activity in the RadioNet FP7 European programme is based on eight
FPGAs interconnected by a high speed transceiver mesh. It provides up to ~4
TMACs with Altera Stratix IV FPGA and 160 Gbps data rate for the input data
stream.
Considering the high in-out data rate in the pre-correlation stages, only
real-time and go-through detectors (i.e. no iterative processing) can be
implemented. In this paper, a real-time and adaptive detection scheme is
described.
An ongoing case study has been set up with the Electronic Multi-Beam Radio
Astronomy Concept (EMBRACE) radio telescope facility at Nan\c{c}ay Observatory.
The objective is to evaluate the performances of this concept in term of
hardware complexity, detection efficiency and additional RFI metadata rate
cost. The UniBoard implementation scheme is described.Comment: 16 pages, 13 figure
The ALICE TPC, a large 3-dimensional tracking device with fast readout for ultra-high multiplicity events
The design, construction, and commissioning of the ALICE Time-Projection
Chamber (TPC) is described. It is the main device for pattern recognition,
tracking, and identification of charged particles in the ALICE experiment at
the CERN LHC. The TPC is cylindrical in shape with a volume close to 90 m^3 and
is operated in a 0.5 T solenoidal magnetic field parallel to its axis.
In this paper we describe in detail the design considerations for this
detector for operation in the extreme multiplicity environment of central
Pb--Pb collisions at LHC energy. The implementation of the resulting
requirements into hardware (field cage, read-out chambers, electronics),
infrastructure (gas and cooling system, laser-calibration system), and software
led to many technical innovations which are described along with a presentation
of all the major components of the detector, as currently realized. We also
report on the performance achieved after completion of the first round of
stand-alone calibration runs and demonstrate results close to those specified
in the TPC Technical Design Report.Comment: 55 pages, 82 figure
Design and management of image processing pipelines within CPS: Acquired experience towards the end of the FitOptiVis ECSEL Project
Cyber-Physical Systems (CPSs) are dynamic and reactive systems interacting with processes, environment and, sometimes, humans. They are often distributed with sensors and actuators, characterized for being smart, adaptive, predictive and react in real-time. Indeed, image- and video-processing pipelines are a prime source for environmental information for systems allowing them to take better decisions according to what they see. Therefore, in FitOptiVis, we are developing novel methods and tools to integrate complex image- and video-processing pipelines. FitOptiVis aims to deliver a reference architecture for describing and optimizing quality and resource management for imaging and video pipelines in CPSs both at design- and run-time. The architecture is concretized in low-power, high-performance, smart components, and in methods and tools for combined design-time and run-time multi-objective optimization and adaptation within system and environment constraints
A Comprehensive Workflow for General-Purpose Neural Modeling with Highly Configurable Neuromorphic Hardware Systems
In this paper we present a methodological framework that meets novel
requirements emerging from upcoming types of accelerated and highly
configurable neuromorphic hardware systems. We describe in detail a device with
45 million programmable and dynamic synapses that is currently under
development, and we sketch the conceptual challenges that arise from taking
this platform into operation. More specifically, we aim at the establishment of
this neuromorphic system as a flexible and neuroscientifically valuable
modeling tool that can be used by non-hardware-experts. We consider various
functional aspects to be crucial for this purpose, and we introduce a
consistent workflow with detailed descriptions of all involved modules that
implement the suggested steps: The integration of the hardware interface into
the simulator-independent model description language PyNN; a fully automated
translation between the PyNN domain and appropriate hardware configurations; an
executable specification of the future neuromorphic system that can be
seamlessly integrated into this biology-to-hardware mapping process as a test
bench for all software layers and possible hardware design modifications; an
evaluation scheme that deploys models from a dedicated benchmark library,
compares the results generated by virtual or prototype hardware devices with
reference software simulations and analyzes the differences. The integration of
these components into one hardware-software workflow provides an ecosystem for
ongoing preparative studies that support the hardware design process and
represents the basis for the maturity of the model-to-hardware mapping
software. The functionality and flexibility of the latter is proven with a
variety of experimental results
A Case for Leveraging 802.11p for Direct Phone-to-Phone Communications
WiFi cannot effectively handle the demands of device-to-device communication between phones, due to insufficient range and poor reliability. We make the case for using IEEE 802.11p DSRC instead, which has been adopted for vehicle-to-vehicle communications, providing lower latency and longer range. We demonstrate a prototype motivated by a novel fabrication process that deposits both III-V and CMOS devices on the same die. In our system prototype, the designed RF front-end is interfaced with a baseband processor on an FPGA, connected to Android phones. It consumes 0.02uJ/bit across 100m assuming free space. Application-level power control dramatically reduces power consumption by 47-56%.Singapore-MIT Alliance for Research and TechnologyAmerican Society for Engineering Education. National Defense Science and Engineering Graduate Fellowshi
Adaptive multispectral GPU accelerated architecture for Earth Observation satellites
In recent years the growth in quantity, diversity and capability of Earth Observation (EO) satellites, has enabled increase’s in the achievable payload data dimensionality and volume. However, the lack of equivalent advancement in downlink technology has resulted in the development of an onboard data bottleneck. This bottleneck must be alleviated in order for EO satellites to continue to efficiently provide high quality and increasing quantities of payload data. This research explores the selection and implementation of state-of-the-art multidimensional image compression algorithms and proposes a new onboard data processing architecture, to help alleviate the bottleneck and increase the data throughput of the platform. The proposed new system is based upon a backplane architecture to provide scalability with different satellite platform sizes and varying mission’s objectives. The heterogeneous nature of the architecture allows benefits of both Field Programmable Gate Array (FPGA) and Graphical Processing Unit (GPU) hardware to be leveraged for maximised data processing throughput
Recommended from our members
FPGA Security Techniques with Applications to Cloud and Multi-Tenant Use Cases
Field programmable gate arrays (FPGAs) are integrated circuits that consist of programmable logic that a user can configure and deploy for applications such as hardware emulation and accelerating high performance computing. In recent years, the emergence of FPGAs in the cloud has led to research on multi-tenant FPGAs. In a multi-tenant scenario, the same FPGA fabric is shared among multiple users, or among multiple untrusting IP cores. Multi-tenancy has economic benefits, largely due to improvements in resource utilization, but also brings new security concerns since the tenants could behave maliciously. Although the tenants sharing an FPGA are logically isolated from each other, they may still have unintended interactions through side channel attacks and fault attacks. In this dissertation, we aim to evaluate security threats and defenses in the multi-tenant FPGA scenario. Firstly, the work in this dissertation studies a true random number generator (TRNG) on cloud FPGAs that is robust against voltage manipulation from co-tenants. The TRNG design is based on harvesting clock jitter using a tunable time-to-digital converter circuit. In accordance with best practices, a stochastic model is built to evaluate the min-entropy of the design, and further validated by NIST entropy assessment test suite and NIST statistical tests. The basic version of the TRNG is extended with a linkable sampling module to increase min-entropy per sample and throughput at a modest resource cost. Then the dissertation analyzes a type of fault attack that can be conducted by one tenant against another in a multi-tenant setting. Specifically, the fault attack is differential fault intensity analysis (DFIA), which is a biased-fault based attack on Advanced Encryption Standard (AES) circuits. Ring oscillators (ROs) are deployed as effective power wasters to cause a supply voltage drop through the shared power distribution network (PDN) of tenants. The attack is highly relevant to multi-tenant scenarios because the attacking tenant can create the voltage drop without physical access, and can precisely control the shape of the voltage drop by adjusting both the number of activated ROs and their duration as required for the attack. The voltage drop will in turn increase the delay in the logic and eventually cause specific timing faults which are analyzed to successfully recover the AES keys. In the last part, we use on-chip voltage sensors to detect the location of a target circuits. The sensing scheme leverages time-to-digital converters (TDCs) as voltage sensors, and a novel differential analysis is applied to the sensor data. In a multi-tenant setting, this method can be used either as part of a defensive scheme to monitor against attacks, or it can be used to probe a system and determine how to effectively target an attack to a particular co-tenant victim
- …