SOTAVerified

Epipolar Attention Field Transformers for Bird's Eye View Semantic Segmentation

2024-12-02Unverified0· sign in to hype

Christian Witte, Jens Behley, Cyrill Stachniss, Marvin Raaijmakers

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Spatial understanding of the semantics of the surroundings is a key capability needed by autonomous cars to enable safe driving decisions. Recently, purely vision-based solutions have gained increasing research interest. In particular, approaches extracting a bird's eye view (BEV) from multiple cameras have demonstrated great performance for spatial understanding. This paper addresses the dependency on learned positional encodings to correlate image and BEV feature map elements for transformer-based methods. We propose leveraging epipolar geometric constraints to model the relationship between cameras and the BEV by Epipolar Attention Fields. They are incorporated into the attention mechanism as a novel attribution term, serving as an alternative to learned positional encodings. Experiments show that our method EAFormer outperforms previous BEV approaches by 2% mIoU for map semantic segmentation and exhibits superior generalization capabilities compared to implicitly learning the camera configuration.

Tasks

Reproductions