Cross-view Transformers for real-time Map-view Semantic Segmentation

CVPR 2022  ·  Brady Zhou, Philipp Krähenbühl ·

We present cross-view transformers, an efficient attention-based model for map-view semantic segmentation from multiple cameras. Our architecture implicitly learns a mapping from individual camera views into a canonical map-view representation using a camera-aware cross-view attention mechanism. Each camera uses positional embeddings that depend on its intrinsic and extrinsic calibration. These embeddings allow a transformer to learn the mapping across different views without ever explicitly modeling it geometrically. The architecture consists of a convolutional image encoder for each view and cross-view transformer layers to infer a map-view semantic segmentation. Our model is simple, easily parallelizable, and runs in real-time. The presented architecture performs at state-of-the-art on the nuScenes dataset, with 4x faster inference speeds. Code is available at https://github.com/bradyz/cross_view_transformers.

PDF Abstract CVPR 2022 PDF CVPR 2022 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Bird's-Eye View Semantic Segmentation nuScenes CVT IoU veh - 224x480 - No vis filter - 100x100 at 0.5 31.4 # 9
IoU veh - 448x800 - No vis filter - 100x100 at 0.5 32.5 # 6
IoU veh - 224x480 - Vis filter. - 100x100 at 0.5 36.0 # 8
IoU veh - 448x800 - Vis filter. - 100x100 at 0.5 37.7 # 6

Methods


No methods listed for this paper. Add relevant methods here