Deep feature selection-and-fusion for RGB-D semantic segmentation

10 May 2021  ·  Yuejiao Su, Yuan Yuan, Zhiyu Jiang ·

Scene depth information can help visual information for more accurate semantic segmentation. However, how to effectively integrate multi-modality information into representative features is still an open problem. Most of the existing work uses DCNNs to implicitly fuse multi-modality information. But as the network deepens, some critical distinguishing features may be lost, which reduces the segmentation performance. This work proposes a unified and efficient feature selectionand-fusion network (FSFNet), which contains a symmetric cross-modality residual fusion module used for explicit fusion of multi-modality information. Besides, the network includes a detailed feature propagation module, which is used to maintain low-level detailed information during the forward process of the network. Compared with the state-of-the-art methods, experimental evaluations demonstrate that the proposed model achieves competitive performance on two public datasets.

PDF Abstract

Datasets


Results from the Paper


Ranked #9 on Semantic Segmentation on SUN-RGBD (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Semantic Segmentation NYU Depth v2 FSFNet Mean IoU 52.0% # 35
Semantic Segmentation SUN-RGBD FSFNet Mean IoU 50.6% # 9

Methods


No methods listed for this paper. Add relevant methods here