24,100 research outputs found

    Robotic Pick-and-Place of Novel Objects in Clutter with Multi-Affordance Grasping and Cross-Domain Image Matching

    Full text link
    This paper presents a robotic pick-and-place system that is capable of grasping and recognizing both known and novel objects in cluttered environments. The key new feature of the system is that it handles a wide range of object categories without needing any task-specific training data for novel objects. To achieve this, it first uses a category-agnostic affordance prediction algorithm to select and execute among four different grasping primitive behaviors. It then recognizes picked objects with a cross-domain image classification framework that matches observed images to product images. Since product images are readily available for a wide range of objects (e.g., from the web), the system works out-of-the-box for novel objects without requiring any additional training data. Exhaustive experimental results demonstrate that our multi-affordance grasping achieves high success rates for a wide variety of objects in clutter, and our recognition algorithm achieves high accuracy for both known and novel grasped objects. The approach was part of the MIT-Princeton Team system that took 1st place in the stowing task at the 2017 Amazon Robotics Challenge. All code, datasets, and pre-trained models are available online at http://arc.cs.princeton.eduComment: Project webpage: http://arc.cs.princeton.edu Summary video: https://youtu.be/6fG7zwGfIk

    SurfNet: Generating 3D shape surfaces using deep residual networks

    Full text link
    3D shape models are naturally parameterized using vertices and faces, \ie, composed of polygons forming a surface. However, current 3D learning paradigms for predictive and generative tasks using convolutional neural networks focus on a voxelized representation of the object. Lifting convolution operators from the traditional 2D to 3D results in high computational overhead with little additional benefit as most of the geometry information is contained on the surface boundary. Here we study the problem of directly generating the 3D shape surface of rigid and non-rigid shapes using deep convolutional neural networks. We develop a procedure to create consistent `geometry images' representing the shape surface of a category of 3D objects. We then use this consistent representation for category-specific shape surface generation from a parametric representation or an image by developing novel extensions of deep residual networks for the task of geometry image generation. Our experiments indicate that our network learns a meaningful representation of shape surfaces allowing it to interpolate between shape orientations and poses, invent new shape surfaces and reconstruct 3D shape surfaces from previously unseen images.Comment: CVPR 2017 pape

    Tightly Coupled GNSS and Vision Navigation for Unmanned Air Vehicle Applications

    Get PDF
    This paper explores the unique benefits that can be obtained from a tight integration of a GNSS sensor and a forward-looking vision sensor. The motivation of this research is the belief that both GNSS and vision will be integral features of future UAV avionics architectures, GNSS for basic aircraft navigation and vision for obstacle-aircraft collision avoidance. The paper will show that utilising basic single-antenna GNSS measurements and observables, along with aircraft information derived from optical flow techniques creates unique synergies. Results of the accuracy of attitude estimates will be presented, based a comprehensive Matlab® Simulink® model which re-creates an optical flow stream based on the flight of an aircraft. This paper establishes the viability of this novel integrated GNSS/Vision approach for use as the complete UAV sensor package, or as a backup sensor for an inertial navigation system

    Deep residual learning in CT physics: scatter correction for spectral CT

    Full text link
    Recently, spectral CT has been drawing a lot of attention in a variety of clinical applications primarily due to its capability of providing quantitative information about material properties. The quantitative integrity of the reconstructed data depends on the accuracy of the data corrections applied to the measurements. Scatter correction is a particularly sensitive correction in spectral CT as it depends on system effects as well as the object being imaged and any residual scatter is amplified during the non-linear material decomposition. An accurate way of removing scatter is subtracting the scatter estimated by Monte Carlo simulation. However, to get sufficiently good scatter estimates, extremely large numbers of photons is required, which may lead to unexpectedly high computational costs. Other approaches model scatter as a convolution operation using kernels derived using empirical methods. These techniques have been found to be insufficient in spectral CT due to their inability to sufficiently capture object dependence. In this work, we develop a deep residual learning framework to address both issues of computation simplicity and object dependency. A deep convolution neural network is trained to determine the scatter distribution from the projection content in training sets. In test cases of a digital anthropomorphic phantom and real water phantom, we demonstrate that with much lower computing costs, the proposed network provides sufficiently accurate scatter estimation
    corecore