334 research outputs found

    Fast vision through frameless event-based sensing and convolutional processing: Application to texture recognition

    Get PDF
    Address-event representation (AER) is an emergent hardware technology which shows a high potential for providing in the near future a solid technological substrate for emulating brain-like processing structures. When used for vision, AER sensors and processors are not restricted to capturing and processing still image frames, as in commercial frame-based video technology, but sense and process visual information in a pixel-level event-based frameless manner. As a result, vision processing is practically simultaneous to vision sensing, since there is no need to wait for sensing full frames. Also, only meaningful information is sensed, communicated, and processed. Of special interest for brain-like vision processing are some already reported AER convolutional chips, which have revealed a very high computational throughput as well as the possibility of assembling large convolutional neural networks in a modular fashion. It is expected that in a near future we may witness the appearance of large scale convolutional neural networks with hundreds or thousands of individual modules. In the meantime, some research is needed to investigate how to assemble and configure such large scale convolutional networks for specific applications. In this paper, we analyze AER spiking convolutional neural networks for texture recognition hardware applications. Based on the performance figures of already available individual AER convolution chips, we emulate large scale networks using a custom made event-based behavioral simulator. We have developed a new event-based processing architecture that emulates with AER hardware Manjunath's frame-based feature recognition software algorithm, and have analyzed its performance using our behavioral simulator. Recognition rate performance is not degraded. However, regarding speed, we show that recognition can be achieved before an equivalent frame is fully sensed and transmitted.Ministerio de Educación y Ciencia TEC-2006-11730-C03-01Junta de Andalucía P06-TIC-01417European Union IST-2001-34124, 21677

    Stacked Cross Validation with Deep Features: A Hybrid Method for Skin Cancer Detection

    Get PDF
    Detection of malignant skin lesions is important for early and accurate diagnosis of skin cancer. In this work, a hybrid method for malignant lesion detection from dermoscopy images is proposed. The method combines the feature extraction process of convolutional neural networks (CNN) with an ensemble learner called stacked cross-validation (CV). The features extracted by three different CNN architectures, namely, ResNet50, Xception, and VGG16 are used for training of four different baseline classifiers, which are support vector machines, k-nearest neighbors, artificial neural networks, and random forests. The stacked outputs of these classifiers are used to train a logistic regression model as a meta-classifier. The performance of the proposed method is compared with the baseline classifiers trained individually as well as AdaBoost classifier, another ensemble learner. Feature extraction with Xception architecture, outperforms all other benchmark models by achieving scores of 0.909, 0.896, 0.886, and 0.917 for accuracy, F1-score, sensitivity, and AUC, respectively

    AUTOMATIC RECOGNITION OF FACIAL EXPRESSION BASED ON COMPUTER VISION

    Full text link

    Efficient Image Processing Via Compressive Sensing Of Integrate-And-Fire Neuronal Network Dynamics

    Get PDF
    Integrate-and-fire (I&F) neuronal networks are ubiquitous in diverse image processing applications, including image segmentation and visual perception. While conventional I&F network image processing requires the number of nodes composing the network to be equal to the number of image pixels driving the network, we determine whether I&F dynamics can accurately transmit image information when there are significantly fewer nodes than network input-signal components. Although compressive sensing (CS) theory facilitates the recovery of images using very few samples through linear signal processing, it does not address whether similar signal recovery techniques facilitate reconstructions through measurement of the nonlinear dynamics of an I&F network. In this paper, we present a new framework for recovering sparse inputs of nonlinear neuronal networks via compressive sensing. By recovering both one-dimensional inputs and two-dimensional images, resembling natural stimuli, we demonstrate that input information can be well-preserved through nonlinear I&F network dynamics even when the number of network-output measurements is significantly smaller than the number of input-signal components. This work suggests an important extension of CS theory potentially useful in improving the processing of medical or natural images through I&F network dynamics and understanding the transmission of stimulus information across the visual system
    corecore