319 research outputs found
Multi-stream CNN based Video Semantic Segmentation for Automated Driving
Majority of semantic segmentation algorithms operate on a single frame even
in the case of videos. In this work, the goal is to exploit temporal
information within the algorithm model for leveraging motion cues and temporal
consistency. We propose two simple high-level architectures based on Recurrent
FCN (RFCN) and Multi-Stream FCN (MSFCN) networks. In case of RFCN, a recurrent
network namely LSTM is inserted between the encoder and decoder. MSFCN combines
the encoders of different frames into a fused encoder via 1x1 channel-wise
convolution. We use a ResNet50 network as the baseline encoder and construct
three networks namely MSFCN of order 2 & 3 and RFCN of order 2. MSFCN-3
produces the best results with an accuracy improvement of 9% and 15% for
Highway and New York-like city scenarios in the SYNTHIA-CVPR'16 dataset using
mean IoU metric. MSFCN-3 also produced 11% and 6% for SegTrack V2 and DAVIS
datasets over the baseline FCN network. We also designed an efficient version
of MSFCN-2 and RFCN-2 using weight sharing among the two encoders. The
efficient MSFCN-2 provided an improvement of 11% and 5% for KITTI and SYNTHIA
with negligible increase in computational complexity compared to the baseline
version.Comment: Accepted for Oral Presentation at VISAPP 201
NeurAll: Towards a Unified Visual Perception Model for Automated Driving
Convolutional Neural Networks (CNNs) are successfully used for the important
automotive visual perception tasks including object recognition, motion and
depth estimation, visual SLAM, etc. However, these tasks are typically
independently explored and modeled. In this paper, we propose a joint
multi-task network design for learning several tasks simultaneously. Our main
motivation is the computational efficiency achieved by sharing the expensive
initial convolutional layers between all tasks. Indeed, the main bottleneck in
automated driving systems is the limited processing power available on
deployment hardware. There is also some evidence for other benefits in
improving accuracy for some tasks and easing development effort. It also offers
scalability to add more tasks leveraging existing features and achieving better
generalization. We survey various CNN based solutions for visual perception
tasks in automated driving. Then we propose a unified CNN model for the
important tasks and discuss several advanced optimization and architecture
design techniques to improve the baseline model. The paper is partly review and
partly positional with demonstration of several preliminary results promising
for future research. We first demonstrate results of multi-stream learning and
auxiliary learning which are important ingredients to scale to a large
multi-task model. Finally, we implement a two-stream three-task network which
performs better in many cases compared to their corresponding single-task
models, while maintaining network size.Comment: Accepted for Oral Presentation at IEEE Intelligent Transportation
Systems Conference (ITSC) 201
Self-Supervised Online Camera Calibration for Automated Driving and Parking Applications
Camera-based perception systems play a central role in modern autonomous
vehicles. These camera based perception algorithms require an accurate
calibration to map the real world distances to image pixels. In practice,
calibration is a laborious procedure requiring specialised data collection and
careful tuning. This process must be repeated whenever the parameters of the
camera change, which can be a frequent occurrence in autonomous vehicles. Hence
there is a need to calibrate at regular intervals to ensure the camera is
accurate. Proposed is a deep learning framework to learn intrinsic and
extrinsic calibration of the camera in real time. The framework is
self-supervised and doesn't require any labelling or supervision to learn the
calibration parameters. The framework learns calibration without the need for
any physical targets or to drive the car on special planar surfaces
Fast and Efficient Scene Categorization for Autonomous Driving using VAEs
Scene categorization is a useful precursor task that provides prior knowledge
for many advanced computer vision tasks with a broad range of applications in
content-based image indexing and retrieval systems. Despite the success of data
driven approaches in the field of computer vision such as object detection,
semantic segmentation, etc., their application in learning high-level features
for scene recognition has not achieved the same level of success. We propose to
generate a fast and efficient intermediate interpretable generalized global
descriptor that captures coarse features from the image and use a
classification head to map the descriptors to 3 scene categories: Rural, Urban
and Suburban. We train a Variational Autoencoder in an unsupervised manner and
map images to a constrained multi-dimensional latent space and use the latent
vectors as compact embeddings that serve as global descriptors for images. The
experimental results evidence that the VAE latent vectors capture coarse
information from the image, supporting their usage as global descriptors. The
proposed global descriptor is very compact with an embedding length of 128,
significantly faster to compute, and is robust to seasonal and illuminational
changes, while capturing sufficient scene information required for scene
categorization.Comment: Published in the 24th Irish Machine Vision and Image Processing
Conference (IMVIP 2022
Effects of low vacuum levels on vacuum dynamics during milking
One of critical points of the milking unit is the short milk tube. Here milk plugs can cause abrupt variations in vacuum which are stressful for the animals. Our trials allowed us to define the effects of the operational vacuum and pulsation on vacuum stability in the short milk tube. Reducing the vacuum from 42 to 28 kPa did not produce appreciable variations in vacuum fluctuation. It was 9.2 kPa for the low vacuum and 9.8 kPa for the standard vacuum. Changing the pulsation rate from 150 to 120 cycles/min did not modify the vacuum stability in the short milk tube. By contrast, raising the pulsation ratio from 50% to 60% significantly increased the amplitude of vacuum fluctuation in the short milk tube
- …
