423 research outputs found

    Fast Back-Projection for Non-Line of Sight Reconstruction

    Get PDF
    Recent works have demonstrated non-line of sight (NLOS) reconstruction by using the time-resolved signal frommultiply scattered light. These works combine ultrafast imaging systems with computation, which back-projects the recorded space-time signal to build a probabilistic map of the hidden geometry. Unfortunately, this computation is slow, becoming a bottleneck as the imaging technology improves. In this work, we propose a new back-projection technique for NLOS reconstruction, which is up to a thousand times faster than previous work, with almost no quality loss. We base on the observation that the hidden geometry probability map can be built as the intersection of the three-bounce space-time manifolds defined by the light illuminating the hidden geometry and the visible point receiving the scattered light from such hidden geometry. This allows us to pose the reconstruction of the hidden geometry as the voxelization of these space-time manifolds, which has lower theoretic complexity and is easily implementable in the GPU. We demonstrate the efficiency and quality of our technique compared against previous methods in both captured and synthetic dat

    Hybrid Bayesian Eigenobjects: Combining Linear Subspace and Deep Network Methods for 3D Robot Vision

    Full text link
    We introduce Hybrid Bayesian Eigenobjects (HBEOs), a novel representation for 3D objects designed to allow a robot to jointly estimate the pose, class, and full 3D geometry of a novel object observed from a single viewpoint in a single practical framework. By combining both linear subspace methods and deep convolutional prediction, HBEOs efficiently learn nonlinear object representations without directly regressing into high-dimensional space. HBEOs also remove the onerous and generally impractical necessity of input data voxelization prior to inference. We experimentally evaluate the suitability of HBEOs to the challenging task of joint pose, class, and shape inference on novel objects and show that, compared to preceding work, HBEOs offer dramatically improved performance in all three tasks along with several orders of magnitude faster runtime performance.Comment: To appear in the International Conference on Intelligent Robots (IROS) - Madrid, 201

    Study and evaluation of correlation techniques for automotive sensor data

    Get PDF
    openI sistemi di guida completamente autonomi sono dotati di avanzati sensori in grado di rilevare e riconoscere rapidamente oggetti nell'ambiente circostante. Tuttavia, la generazione di enormi volumi di dati da parte di questi sensori può rappresentare una sfida considerevole per le tecnologie di comunicazione standard. In questa tesi, ci proponiamo di esaminare e convalidare diversi metodi per valutare la correlazione dei dati automobilistici utilizzando un dataset sintetico, al fine di determinare il modo più efficace per trasmettere tali dati. La nostra ricerca esplora vari metodi di correlazione, tra cui il coefficiente di correlazione di Pearson, la Chamfer Distance, l'algoritmo Iterative Closest Point (ICP) e la Normal Distribution Transform (NDT). Inoltre, sviluppiamo una nuova funzione di punteggio che combina la Chamfer Distance e la correlazione al fine di individuare cambiamenti significativi nei dati automobilistici. La nostra indagine si basa su un dataset sintetico e considera contesti sia statici che dinamici. I nostri risultati mettono in evidenza l'importante interazione tra correlazione e Chamfer Distance in ambienti dinamici e dimostrano che i dati automobilistici sono intrinsecamente correlati, pur richiedendo una certa ridondanza per garantire l'accuratezza, specialmente nelle aree critiche in cui sono essenziali i requisiti di sicurezza.Fully autonomous driving systems are equipped with sensors able to guarantee fast detection and recognition of sensitive objects in the environment. However, the resulting huge volumes of data generated from those sensors may be challenging to handle for standard communication technologies. Along these lines, in this thesis we test and validate different methods to evaluate correlation of automotive data on a synthetic dataset, to decide the most convenient way of transmitting data. The study investigates correlation methods such as Pearson’s correlation coefficient, the Chamfer Distance, the Iterative Closest Point (ICP), and the Normal Distribution Transform (NDT) algorithms. We also develop a new score function combining Chamfer Distance and correlation to detect significant changes in the automotive data. The research utilizes a synthetic dataset and examines both static and dynamic contexts. Our findings highlight the interplay between correlation and Chamfer Distance in dynamic environments, and demonstrate that automotive data are highly correlated, even though redundancy is needed to guarantee accuracy, especially in critical areas where safety requirements are particularly critical

    Unsupervised Object-Centric Voxelization for Dynamic Scene Understanding

    Full text link
    Understanding the compositional dynamics of multiple objects in unsupervised visual environments is challenging, and existing object-centric representation learning methods often ignore 3D consistency in scene decomposition. We propose DynaVol, an inverse graphics approach that learns object-centric volumetric representations in a neural rendering framework. DynaVol maintains time-varying 3D voxel grids that explicitly represent the probability of each spatial location belonging to different objects, and decouple temporal dynamics and spatial information by learning a canonical-space deformation field. To optimize the volumetric features, we embed them into a fully differentiable neural network, binding them to object-centric global features and then driving a compositional NeRF for scene reconstruction. DynaVol outperforms existing methods in novel view synthesis and unsupervised scene decomposition and allows for the editing of dynamic scenes, such as adding, deleting, replacing objects, and modifying their trajectories

    Dense 3D Object Reconstruction from a Single Depth View

    Get PDF
    In this paper, we propose a novel approach, 3D-RecGAN++, which reconstructs the complete 3D structure of a given object from a single arbitrary depth view using generative adversarial networks. Unlike existing work which typically requires multiple views of the same object or class labels to recover the full 3D geometry, the proposed 3D-RecGAN++ only takes the voxel grid representation of a depth view of the object as input, and is able to generate the complete 3D occupancy grid with a high resolution of 256^3 by recovering the occluded/missing regions. The key idea is to combine the generative capabilities of autoencoders and the conditional Generative Adversarial Networks (GAN) framework, to infer accurate and fine-grained 3D structures of objects in high-dimensional voxel space. Extensive experiments on large synthetic datasets and real-world Kinect datasets show that the proposed 3D-RecGAN++ significantly outperforms the state of the art in single view 3D object reconstruction, and is able to reconstruct unseen types of objects.Comment: TPAMI 2018. Code and data are available at: https://github.com/Yang7879/3D-RecGAN-extended. This article extends from arXiv:1708.0796

    3D-PhysNet: Learning the Intuitive Physics of Non-Rigid Object Deformations

    Full text link
    The ability to interact and understand the environment is a fundamental prerequisite for a wide range of applications from robotics to augmented reality. In particular, predicting how deformable objects will react to applied forces in real time is a significant challenge. This is further confounded by the fact that shape information about encountered objects in the real world is often impaired by occlusions, noise and missing regions e.g. a robot manipulating an object will only be able to observe a partial view of the entire solid. In this work we present a framework, 3D-PhysNet, which is able to predict how a three-dimensional solid will deform under an applied force using intuitive physics modelling. In particular, we propose a new method to encode the physical properties of the material and the applied force, enabling generalisation over materials. The key is to combine deep variational autoencoders with adversarial training, conditioned on the applied force and the material properties. We further propose a cascaded architecture that takes a single 2.5D depth view of the object and predicts its deformation. Training data is provided by a physics simulator. The network is fast enough to be used in real-time applications from partial views. Experimental results show the viability and the generalisation properties of the proposed architecture.Comment: in IJCAI 201
    • …
    corecore