This paper presents a system for improving the robustness of LiDAR lateral
localisation systems. This is made possible by including detections of road
boundaries which are invisible to the sensor (due to occlusion, e.g. traffic)
but can be located by our Occluded Road Boundary Inference Deep Neural Network.
We show an example application in which fusion of a camera stream is used to
initialise the lateral localisation. We demonstrate over four driven forays
through central Oxford - totalling 40 km of driving - a gain in performance
that inferring of occluded road boundaries brings.Comment: accepted for publication at the IEEE/ION Position, Location and
Navigation Symposium (PLANS) 202