16,512 research outputs found
Complete Solution for Vehicle Re-ID in Surround-view Camera System
Vehicle re-identification (Re-ID) is a critical component of the autonomous
driving perception system, and research in this area has accelerated in recent
years. However, there is yet no perfect solution to the vehicle
re-identification issue associated with the car's surround-view camera system.
Our analysis identifies two significant issues in the aforementioned scenario:
i) It is difficult to identify the same vehicle in many picture frames due to
the unique construction of the fisheye camera. ii) The appearance of the same
vehicle when seen via the surround vision system's several cameras is rather
different. To overcome these issues, we suggest an integrative vehicle Re-ID
solution method. On the one hand, we provide a technique for determining the
consistency of the tracking box drift with respect to the target. On the other
hand, we combine a Re-ID network based on the attention mechanism with spatial
limitations to increase performance in situations involving multiple cameras.
Finally, our approach combines state-of-the-art accuracy with real-time
performance. We will soon make the source code and annotated fisheye dataset
available.Comment: 11 pages, 10 figures. arXiv admin note: substantial text overlap with
arXiv:2006.1650
Composite video and graphics display for camera viewing systems in robotics and teleoperation
A system for real-time video image display for robotics or remote-vehicle teleoperation is described that has at least one robot arm or remotely operated vehicle controlled by an operator through hand-controllers, and one or more television cameras and optional lighting element. The system has at least one television monitor for display of a television image from a selected camera and the ability to select one of the cameras for image display. Graphics are generated with icons of cameras and lighting elements for display surrounding the television image to provide the operator information on: the location and orientation of each camera and lighting element; the region of illumination of each lighting element; the viewed region and range of focus of each camera; which camera is currently selected for image display for each monitor; and when the controller coordinate for said robot arms or remotely operated vehicles have been transformed to correspond to coordinates of a selected or nonselected camera
End-to-End Learning of Driving Models with Surround-View Cameras and Route Planners
For human drivers, having rear and side-view mirrors is vital for safe
driving. They deliver a more complete view of what is happening around the car.
Human drivers also heavily exploit their mental map for navigation.
Nonetheless, several methods have been published that learn driving models with
only a front-facing camera and without a route planner. This lack of
information renders the self-driving task quite intractable. We investigate the
problem in a more realistic setting, which consists of a surround-view camera
system with eight cameras, a route planner, and a CAN bus reader. In
particular, we develop a sensor setup that provides data for a 360-degree view
of the area surrounding the vehicle, the driving route to the destination, and
low-level driving maneuvers (e.g. steering angle and speed) by human drivers.
With such a sensor setup we collect a new driving dataset, covering diverse
driving scenarios and varying weather/illumination conditions. Finally, we
learn a novel driving model by integrating information from the surround-view
cameras and the route planner. Two route planners are exploited: 1) by
representing the planned routes on OpenStreetMap as a stack of GPS coordinates,
and 2) by rendering the planned routes on TomTom Go Mobile and recording the
progression into a video. Our experiments show that: 1) 360-degree
surround-view cameras help avoid failures made with a single front-view camera,
in particular for city driving and intersection scenarios; and 2) route
planners help the driving task significantly, especially for steering angle
prediction.Comment: to be published at ECCV 201
- …