Vehicle visual perception is essential for safe autonomous driving, especially in challenging low-visibility conditions. Polarimetric imaging has shown enhanced perception by improving target-background contrast and reducing glare. However, current research in polarimetric imaging for autonomous driving largely focuses on singular use of polarimetric features. Exploiting polarimetric information to enhance vehicle visual perception in low-visibility scenarios remains a critical challenge. This paper addresses this challenge by integrating multiple polarimetric parameter characterization with a deep learning model for semantic segmentation visual perception task, which is called TransWNet. The model combines Convolutional Neural Networks (CNN) and Transformer architectures, extracting features and contextual information of both Degree of Polarization (DoP) and Angle of Polarization (AoP) comprehensively during the encoding phase. In the decoding phase, it incorporates skip connections to effectively retain multimodal polarimetric information across deep and shallow features, generating output through feature fusion. To the best of our knowledge, this is the first work that jointly exploits DoP and AoP for vehicle scene understanding in degraded visibility. Experimental results demonstrate that TransWNet, by effectively leveraging multimodal polarimetric information, achieves significantly better performance in semantic segmentation of low-visibility traffic scenes, with marked improvements in mIoU, mPA, and Accuracy over all single-feature baselines. Compared with the baseline method, TransWNet improves the Accuracy by 3.44%
Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.