81 research outputs found

    An orientation selective 2D AER transceiver

    Get PDF
    This paper describes an address event representation (AER) transceiver chip that accepts 2D images and produces 2D output images equal to the input filtered by even and odd symmetric orientation selective spatial filters. Both input and output are encoded as spike trains using a differential ON/OFF representation, conserving energy and AER bandwidth. The spatial filtering is performed by symmetric analog circuits that operate on input currents obtained by integrating the input spike trains, and which preserve the ON/OFF representation. This chip is a key component of a multi-chip system we are constructing that is inspired by the visual cortex. We present measured results from a 32 x 64 pixel prototype, which was fabricated in the TSMC0.25 μm process on a 3.84mm by 2.54mm die. Quiescent power dissipation was 3mW

    AER Building Blocks for Multi-Layer Multi-Chip Neuromorphic Vision Systems

    Get PDF
    A 5-layer neuromorphic vision processor whose components communicate spike events asychronously using the address-eventrepresentation (AER) is demonstrated. The system includes a retina chip, two convolution chips, a 2D winner-take-all chip, a delay line chip, a learning classifier chip, and a set of PCBs for computer interfacing and address space remappings. The components use a mixture of analog and digital computation and will learn to classify trajectories of a moving object. A complete experimental setup and measurements results are shown.Unión Europea IST-2001-34124 (CAVIAR)Ministerio de Ciencia y Tecnología TIC-2003-08164-C0

    An ON-OFF orientation selective address event representation image transceiver chip

    Get PDF
    This paper describes the electronic implementation of a four-layer cellular neural network architecture implementing two components of a functional model of neurons in the visual cortex: linear orientation selective filtering and half wave rectification. Separate ON and OFF layers represent the positive and negative outputs of two-phase quadrature Gabor-type filters, whose orientation and spatial-frequency tunings are electronically adjustable. To enable the construction of a multichip network to extract different orientations in parallel, the chip includes an address event representation (AER) transceiver that accepts and produces two-dimensional images that are rate encoded as spike trains. It also includes routing circuitry that facilitates point-to-point signal fan in and fan out. We present measured results from a 32 x 64 pixel prototype, which was fabricated in the TSMC0.25-μm process on a 3.84 by 2.54 mm die. Quiescent power dissipation is 3 mW and is determined primarily by the spike activity on the AER bus. Settling times are on the order of a few milliseconds. In comparison with a two-layer network implementing the same filters, this network results in a more symmetric circuit design with lower quiescent power dissipation, albeit at the expense of twice as many transistors

    Real-time Neuromorphic Visual Pre-Processing and Dynamic Saliency

    Get PDF
    The human brain is by far the most computationally complex, efficient, and reliable computing system operating under such low-power, small-size, and light-weight specifications. Within the field of neuromorphic engineering, we seek to design systems with facsimiles to that of the human brain with means to reach its desirable properties. In this doctoral work, the focus is within the realm of vision, specifically visual saliency and related visual tasks with bio-inspired, real-time processing. The human visual system, from the retina through the visual cortical hierarchy, is responsible for extracting visual information and processing this information, forming our visual perception. This visual information is transmitted through these various layers of the visual system via spikes (or action potentials), representing information in the temporal domain. The objective is to exploit this neurological communication protocol and functionality within the systems we design. This approach is essential for the advancement of autonomous, mobile agents (i.e. drones/MAVs, cars) which must perform visual tasks under size and power constraints in which traditional CPU or GPU implementations to not suffice. Although the high-level objective is to design a complete visual processor with direct physical and functional correlates to the human visual system, we focus on three specific tasks. The first focus of this thesis is the integration of motion into a biologically-plausible proto-object-based visual saliency model. Laurent Itti, one of the pioneers in the field, defines visual saliency as ``the distinct subjective perceptual quality which makes some items in the world stand out from their neighbors and immediately grab our attention.'' From humans to insects, visual saliency is important for the extraction of only interesting regions of visual stimuli for further processing. Prior to this doctoral work, Russel et al. \cite{russell2014model} designed a model of proto-object-based visual saliency with biological correlates. This model was designed for computing saliency only on static images. However, motion is a naturally occurring phenomena that plays an essential role in both human and animal visual processing. Henceforth, the most ideal model of visual saliency should consider motion that may be exhibited within the visual scene. In this work a novel dynamic proto-object-based visual saliency is described which extends the Russel et. al. saliency model to consider not only static, but also temporal information. This model was validated by using metrics for determining how accurate the model is in predicting human eye fixations and saccades on a public dataset of videos with attached eye tracking data. This model outperformed other state-of-the-art visual saliency models in computing dynamic visual saliency. Such a model that can accurately predict where humans look, can serve as a front-end component to other visual processors performing tasks such as object detection and recognition, or object tracking. In doing so it can reduce throughput and increase processing speed for such tasks. Furthermore, it has more obvious applications in artificial intelligence in mimicking the functionality of the human visual system. The second focus of this thesis is the implementation of this visual saliency model on an FPGA (Field Programmable Gate Array) for real-time processing. Initially, this model was designed within MATLAB, a software-based approach running on a CPU, which limits the processing speed and consumes unnecessary amounts of power due to overhead. This is detrimental for integration with an autonomous, mobile system which must operate in real-time. This novel FPGA implementation allows for a low-power, high-speed approach to computing visual saliency. There are a few existing FPGA-based implementations of visual saliency, and of those, none are based on the notion of proto-objects. This work presents the first, to our knowledge, FPGA implementation of an object-based visual saliency model. Such an FPGA implementation allows for the low-power, light-weight, and small-size specifications that we seek within the field of neuromorphic engineering. For validating the FPGA model, the same metrics are used for determining the extent to which it predicts human eye saccades and fixations. We compare this hardware implementation to the software model for validation. The third focus of this thesis is the design of a generic neuromorphic platform both on FPGA and VLSI (Very-Large-Scale-Integration) technology for performing visual tasks, including those necessary in the computation of the visual saliency. Visual processing tasks such as image filtering and image dewarping are demonstrated via this novel neuromorphic technology consisting of an array of hardware-based generalized integrate-and-fire neurons. It allows the visual saliency model's computation to be offloaded onto this hardware-based architecture. We first demonstrate an emulation of this neuromorphic system on FPGA demonstrating its capability of dewarping and filtering tasks as well as integration with a neuromorphic camera called the ATIS (Asynchronous Time-based Image Sensor). We then demonstrate the neuromorphic platform implemented in CMOS technology, specifically designed for low-mismatch, high-density, and low-power. Such a VLSI technology-based platform further bridges the gap between engineering and biology and moves us closer towards developing a complete neuromorphic visual processor

    Event-based neuromorphic stereo vision

    Full text link

    Institute of Safety Research, Annual Report 1995

    Get PDF
    The report gives an overview on the scientific work of the Institute of Safety Research in 1995

    Sensor network localization based on natural phenomena

    Get PDF
    Thesis (M. Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2006.Includes bibliographical references (p. 107-116).Autonomous localization is crucial for many sensor network applications. The goal of this thesis is to develop a distributed localization algorithm for the PLUG indoor sensor network by analyzing sound and light sensory data from naturally occurring background phenomena as well as synthesized emulations of background transients. Our approach has two main phases: passive and active. The system enters an active mode when, its sensed region stays relatively silent and stable, hence assumed to be unoccupied; otherwise, it stays in the passive mode. In the passive mode, each node looks for sonic transients and compares the timing of its highest sound peak to that of synchronized sound peaks from other nodes in its neighborhood in order to estimate its distance. Passive ranging achieved 50.96cm error and simulated passive localization achieved 103.06cm error with a typical node-spacing of 2m. In addition, the system exploits background transients based on light sensory data to determine room boundaries. In the active mode, each node occasionally generates recorded mimics of natural sonic transients, like pencils dropping or water glasses clinking and manipulates an attached light source. Active acoustic ranging achieved 2.1cm error and simulated active localization achieved 7.97cm error with a typical node-spacing of 2m. In addition, passive location estimation in a real deployment is found to converge as more sensory data is available; range resolutions of 2.5m and localization errors of 20.3cm were obtained after running in passive mode for 20 hours in 7m by 5m dorm hallway. The main features of author's approach are its distributed properties, the lack of any heavy infrastructure, its unobtrusive exploitation of multi-sensory background phenomena, and in active mode, making the sound signal between nodes unobtrusive by mimicking the natural sounds.by Daniel Sang Kim.M.Eng
    • …
    corecore