15 research outputs found
Ecological active vision: four bio-inspired principles to integrate bottom-up and adaptive top-down attention tested with a simple camera-arm robot
Vision gives primates a wealth of information useful to manipulate the environment, but at the same time it can easily overwhelm their computational resources. Active vision is a key solution found by nature to solve this problem: a limited fovea actively displaced in space to collect only relevant information. Here we highlight that in ecological conditions this solution encounters four problems: 1) the agent needs to learn where to look based on its goals; 2) manipulation causes learning feedback in areas of space possibly outside the attention focus; 3) good visual actions are needed to guide manipulation actions, but only these can generate learning feedback; and 4) a limited fovea causes aliasing problems. We then propose a computational architecture ("BITPIC") to overcome the four problems, integrating four bioinspired key ingredients: 1) reinforcement-learning fovea-based top-down attention; 2) a strong vision-manipulation coupling; 3) bottom-up periphery-based attention; and 4) a novel action-oriented memory. The system is tested with a simple simulated camera-arm robot solving a class of search-and-reach tasks involving color-blob "objects." The results show that the architecture solves the problems, and hence the tasks, very ef?ciently, and highlight how the architecture principles can contribute to a full exploitation of the advantages of active vision in ecological conditions
Improved fully convolutional network with conditional random field for building extraction
Dissertation submitted in partial fulfilment of the requirements for the degree of Master of Science in Geospatial TechnologiesBuilding extraction from remotely sensed imagery plays an important role in urban planning,
disaster management, navigation, updating geographic databases and several other geospatial
applications. Several published contributions are dedicated to the applications of Deep Convolutional
Neural Network (DCNN) for building extraction using aerial/satellite imagery exists;
however, in all these contributions a good accuracy is always paid at the price of extremely
complex and large network architectures. In this paper, we present an enhanced Fully Convolutional
Network (FCN) framework especially molded for building extraction of remotely sensed
images by applying Conditional Random Field (CRF). The main purpose here is to propose
a framework which balances maximum accuracy with less network complexity. The modern
activation function called Exponential Linear Unit (ELU) is applied to improve the performance
of the Fully Convolutional Network (FCN), resulting in more, yet accurate building prediction. To
further reduce the noise (false classified buildings) and to sharpen the boundary of the buildings,
a post processing CRF is added at the end of the adopted Convolutional Neural Network (CNN)
framework. The experiments were conducted on Massachusetts building aerial imagery. The
results show that our proposed framework outperformed FCN baseline, which is the existing
baseline framework for semantic segmentation, in term of performance measure, the F1-score
and Intersection Over Union (IoU) measure. Additionally, the proposed method stood superior to
the pre-existing classifier for building extraction using the same dataset in terms of performance
measure and network complexity at once