41 research outputs found
Recurrent Pixel Embedding for Instance Grouping
We introduce a differentiable, end-to-end trainable framework for solving
pixel-level grouping problems such as instance segmentation consisting of two
novel components. First, we regress pixels into a hyper-spherical embedding
space so that pixels from the same group have high cosine similarity while
those from different groups have similarity below a specified margin. We
analyze the choice of embedding dimension and margin, relating them to
theoretical results on the problem of distributing points uniformly on the
sphere. Second, to group instances, we utilize a variant of mean-shift
clustering, implemented as a recurrent neural network parameterized by kernel
bandwidth. This recurrent grouping module is differentiable, enjoys convergent
dynamics and probabilistic interpretability. Backpropagating the group-weighted
loss through this module allows learning to focus on only correcting embedding
errors that won't be resolved during subsequent clustering. Our framework,
while conceptually simple and theoretically abundant, is also practically
effective and computationally efficient. We demonstrate substantial
improvements over state-of-the-art instance segmentation for object proposal
generation, as well as demonstrating the benefits of grouping loss for
classification tasks such as boundary detection and semantic segmentation
3D-BEVIS: Bird's-Eye-View Instance Segmentation
Recent deep learning models achieve impressive results on 3D scene analysis
tasks by operating directly on unstructured point clouds. A lot of progress was
made in the field of object classification and semantic segmentation. However,
the task of instance segmentation is less explored. In this work, we present
3D-BEVIS, a deep learning framework for 3D semantic instance segmentation on
point clouds. Following the idea of previous proposal-free instance
segmentation approaches, our model learns a feature embedding and groups the
obtained feature space into semantic instances. Current point-based methods
scale linearly with the number of points by processing local sub-parts of a
scene individually. However, to perform instance segmentation by clustering,
globally consistent features are required. Therefore, we propose to combine
local point geometry with global context information from an intermediate
bird's-eye view representation.Comment: camera-ready version for GCPR '1
Collision Avoidance by Identifying Risks for Detected Objects in Autonomous Vehicles
We propose a system which will detect objects onour roads, estimate the distance of these object from the cameraand alert the driver if this distance is equal or less than thethreshold value(02meters),and assist the driver and alert him assoon as possible in order for him to take appropriate actions assoon as possible which can avoid any collision or significantlyreduce it. We plan to use state of the arts object detection modelslike YOLO to identify the target object classes and use depthmaps from monocular camera to be give an accurate estimate ofthe distance of the detected object from the camera. one majorrequirement of this system is the real-time behaviour and a highaccuracy for the detected and estimated distance, A secondrequirement is to make the system cheap and easy useablecomparatively to the other existing methods. That is why wedecided to use monocular camera images and depth maps whichmakes the solution cheap and innovative. This project(prototype) provide room for bigger and more complete projectwhich will contribute to the creation of tool which can save livesand improve security on our road