4D-Former: Multimodal 4D Panoptic Segmentation

2 Nov 2023  ·  Ali Athar, Enxu Li, Sergio Casas, Raquel Urtasun ·

4D panoptic segmentation is a challenging but practically useful task that requires every point in a LiDAR point-cloud sequence to be assigned a semantic class label, and individual objects to be segmented and tracked over time. Existing approaches utilize only LiDAR inputs which convey limited information in regions with point sparsity. This problem can, however, be mitigated by utilizing RGB camera images which offer appearance-based information that can reinforce the geometry-based LiDAR features. Motivated by this, we propose 4D-Former: a novel method for 4D panoptic segmentation which leverages both LiDAR and image modalities, and predicts semantic masks as well as temporally consistent object masks for the input point-cloud sequence. We encode semantic classes and objects using a set of concise queries which absorb feature information from both data modalities. Additionally, we propose a learned mechanism to associate object tracks over time which reasons over both appearance and spatial location. We apply 4D-Former to the nuScenes and SemanticKITTI datasets where it achieves state-of-the-art results.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Panoptic Tracking Panoptic nuScenes test 4D-Former PAT 79.4 # 1
LSTQ 78.2 # 1
PTQ 75.5 # 1
Panoptic Tracking Panoptic nuScenes val 4D-Former PAT 78.3 # 1
LSTQ 76.4 # 1
PTQ 75.2 # 1

Methods


No methods listed for this paper. Add relevant methods here