41 research outputs found

    Recurrent Pixel Embedding for Instance Grouping

    Full text link
    We introduce a differentiable, end-to-end trainable framework for solving pixel-level grouping problems such as instance segmentation consisting of two novel components. First, we regress pixels into a hyper-spherical embedding space so that pixels from the same group have high cosine similarity while those from different groups have similarity below a specified margin. We analyze the choice of embedding dimension and margin, relating them to theoretical results on the problem of distributing points uniformly on the sphere. Second, to group instances, we utilize a variant of mean-shift clustering, implemented as a recurrent neural network parameterized by kernel bandwidth. This recurrent grouping module is differentiable, enjoys convergent dynamics and probabilistic interpretability. Backpropagating the group-weighted loss through this module allows learning to focus on only correcting embedding errors that won't be resolved during subsequent clustering. Our framework, while conceptually simple and theoretically abundant, is also practically effective and computationally efficient. We demonstrate substantial improvements over state-of-the-art instance segmentation for object proposal generation, as well as demonstrating the benefits of grouping loss for classification tasks such as boundary detection and semantic segmentation

    3D-BEVIS: Bird's-Eye-View Instance Segmentation

    Full text link
    Recent deep learning models achieve impressive results on 3D scene analysis tasks by operating directly on unstructured point clouds. A lot of progress was made in the field of object classification and semantic segmentation. However, the task of instance segmentation is less explored. In this work, we present 3D-BEVIS, a deep learning framework for 3D semantic instance segmentation on point clouds. Following the idea of previous proposal-free instance segmentation approaches, our model learns a feature embedding and groups the obtained feature space into semantic instances. Current point-based methods scale linearly with the number of points by processing local sub-parts of a scene individually. However, to perform instance segmentation by clustering, globally consistent features are required. Therefore, we propose to combine local point geometry with global context information from an intermediate bird's-eye view representation.Comment: camera-ready version for GCPR '1

    Collision Avoidance by Identifying Risks for Detected Objects in Autonomous Vehicles

    Get PDF
    We propose a system which will detect objects onour roads, estimate the distance of these object from the cameraand alert the driver if this distance is equal or less than thethreshold value(02meters),and assist the driver and alert him assoon as possible in order for him to take appropriate actions assoon as possible which can avoid any collision or significantlyreduce it. We plan to use state of the arts object detection modelslike YOLO to identify the target object classes and use depthmaps from monocular camera to be give an accurate estimate ofthe distance of the detected object from the camera. one majorrequirement of this system is the real-time behaviour and a highaccuracy for the detected and estimated distance, A secondrequirement is to make the system cheap and easy useablecomparatively to the other existing methods. That is why wedecided to use monocular camera images and depth maps whichmakes the solution cheap and innovative. This project(prototype) provide room for bigger and more complete projectwhich will contribute to the creation of tool which can save livesand improve security on our road
    corecore