6,776 research outputs found
An initial performance review of software components for a heterogeneous computing platform
The design of embedded systems is a complex activity that involves a lot of
decisions. With high performance demands of present day usage scenarios and
software, they often involve energy hungry state-of-the-art computing units.
While focusing on power consumption of computing units, the physical properties
of software are often ignored. Recently, there has been a growing interest to
quantify and model the physical footprint of software (e.g. consumed power,
generated heat, execution time, etc.), and a component based approach
facilitates methods for describing such properties. Based on these, software
architects can make energy-efficient software design solutions. This paper
presents power consumption and execution time profiling of a component software
that can be allocated on heterogeneous computing units (CPU, GPU, FPGA) of a
tracked robot
PCA-RECT: An Energy-efficient Object Detection Approach for Event Cameras
We present the first purely event-based, energy-efficient approach for object
detection and categorization using an event camera. Compared to traditional
frame-based cameras, choosing event cameras results in high temporal resolution
(order of microseconds), low power consumption (few hundred mW) and wide
dynamic range (120 dB) as attractive properties. However, event-based object
recognition systems are far behind their frame-based counterparts in terms of
accuracy. To this end, this paper presents an event-based feature extraction
method devised by accumulating local activity across the image frame and then
applying principal component analysis (PCA) to the normalized neighborhood
region. Subsequently, we propose a backtracking-free k-d tree mechanism for
efficient feature matching by taking advantage of the low-dimensionality of the
feature representation. Additionally, the proposed k-d tree mechanism allows
for feature selection to obtain a lower-dimensional dictionary representation
when hardware resources are limited to implement dimensionality reduction.
Consequently, the proposed system can be realized on a field-programmable gate
array (FPGA) device leading to high performance over resource ratio. The
proposed system is tested on real-world event-based datasets for object
categorization, showing superior classification performance and relevance to
state-of-the-art algorithms. Additionally, we verified the object detection
method and real-time FPGA performance in lab settings under non-controlled
illumination conditions with limited training data and ground truth
annotations.Comment: Accepted in ACCV 2018 Workshops, to appea
A modified neural network model for Lobula Giant Movement Detector with additional depth movement feature
The Lobula Giant Movement Detector (LGMD) is a wide-field visual neuron that is located in the Lobula layer of the Locust nervous system. The LGMD increases its firing rate in response to both the velocity of the approaching object and its proximity. It has been found that it can respond to looming stimuli very quickly and can trigger avoidance reactions whenever a rapidly approaching object is detected. It has been successfully applied in visual collision avoidance systems for vehicles and robots. This paper proposes a modified LGMD model that provides additional movement depth direction information. The proposed model retains the simplicity of the previous neural network model, adding only a few new cells. It has been tested on both simulated and recorded video data sets. The experimental results shows that the modified model can very efficiently provide stable information on the depth direction of movement
A USB3.0 FPGA Event-based Filtering and Tracking Framework for Dynamic Vision Sensors
Dynamic vision sensors (DVS) are frame-free sensors
with an asynchronous variable-rate output that is ideal for hard
real-time dynamic vision applications under power and latency
constraints. Post-processing of the digital sensor output can
reduce sensor noise, extract low level features, and track objects
using simple algorithms that have previously been implemented
in software. In this paper we present an FPGA-based framework
for event-based processing that allows uncorrelated-event noise
removal and real-time tracking of multiple objects, with dynamic
capabilities to adapt itself to fast or slow and large or small
objects. This framework uses a new hardware platform based on
a Lattice FPGA which filters the sensor output and which then
transmits the results through a super-speed Cypress FX3 USB
microcontroller interface to a host computer. The packets of
events and timestamps are transmitted to the host computer at
rates of 10 Mega events per second. Experimental results are
presented that demonstrate a low latency of 10us for tracking
and computing the center of mass of a detected object.Ministerio de Economía y Competitividad TEC2012-37868-C04-0
LEGaTO: first steps towards energy-efficient toolset for heterogeneous computing
LEGaTO is a three-year EU H2020 project which started in December 2017. The LEGaTO project will leverage task-based programming models to provide a software ecosystem for Made-in-Europe heterogeneous hardware composed of CPUs, GPUs, FPGAs and dataflow engines. The aim is to attain one order of magnitude energy savings from the edge to the converged cloud/HPC.Peer ReviewedPostprint (author's final draft
An Approach to Distance Estimation with Stereo Vision Using Address-Event-Representation
Image processing in digital computer systems usually considers the
visual information as a sequence of frames. These frames are from cameras that
capture reality for a short period of time. They are renewed and transmitted at a
rate of 25-30 fps (typical real-time scenario). Digital video processing has to
process each frame in order to obtain a result or detect a feature. In stereo
vision, existing algorithms used for distance estimation use frames from two
digital cameras and process them pixel by pixel to obtain similarities and
differences from both frames; after that, depending on the scene and the
features extracted, an estimate of the distance of the different objects of the
scene is calculated. Spike-based processing is a relatively new approach that
implements the processing by manipulating spikes one by one at the time they
are transmitted, like a human brain. The mammal nervous system is able to
solve much more complex problems, such as visual recognition by
manipulating neuron spikes. The spike-based philosophy for visual information
processing based on the neuro-inspired Address-Event-Representation (AER) is
achieving nowadays very high performances. In this work we propose a two-
DVS-retina system, composed of other elements in a chain, which allow us to
obtain a distance estimation of the moving objects in a close environment. We
will analyze each element of this chain and propose a Multi Hold&Fire
algorithm that obtains the differences between both retinas.Ministerio de Ciencia e Innovación TEC2009-10639-C04-0
Live Demonstration: On the distance estimation of moving targets with a Stereo-Vision AER system
Distance calculation is always one of the most
important goals in a digital stereoscopic vision system. In an
AER system this goal is very important too, but it cannot be
calculated as accurately as we would like. This demonstration
shows a first approximation in this field, using a disparity
algorithm between both retinas. The system can make a distance
approach about a moving object, more specifically, a qualitative
estimation. Taking into account the stereo vision system
features, the previous retina positioning and the very important
Hold&Fire building block, we are able to make a correlation
between the spike rate of the disparity and the distance.Ministerio de Ciencia e Innovación TEC2009-10639-C04-0
- …