Learning to Adapt Structured Output Space for Semantic Segmentation

Convolutional neural network-based approaches for semantic segmentation rely on supervision with pixel-level ground truth, but may not generalize well to unseen image domains. As the labeling process is tedious and labor intensive, developing algorithms that can adapt source ground truth labels to the target domain is of great interest. In this paper, we propose an adversarial learning method for domain adaptation in the context of semantic segmentation. Considering semantic segmentations as structured outputs that contain spatial similarities between the source and target domains, we adopt adversarial learning in the output space. To further enhance the adapted model, we construct a multi-level adversarial network to effectively perform output space domain adaptation at different feature levels. Extensive experiments and ablation study are conducted under various domain adaptation settings, including synthetic-to-real and cross-city scenarios. We show that the proposed method performs favorably against the state-of-the-art methods in terms of accuracy and visual quality.

PDF Abstract CVPR 2018 PDF CVPR 2018 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Synthetic-to-Real Translation GTAV-to-Cityscapes Labels AdaptSegNet(multi-level) mIoU 42.4 # 62
Domain Adaptation Synscapes-to-Cityscapes AdaptSegNet mIoU 52.7 # 3
Synthetic-to-Real Translation SYNTHIA-to-Cityscapes AdaptSegNet(Multi-level) MIoU (13 classes) 46.7 # 33
Image-to-Image Translation SYNTHIA-to-Cityscapes Multi-level Adaptation mIoU (13 classes) 46.7 # 21
Image-to-Image Translation SYNTHIA-to-Cityscapes Single-level Adaptation mIoU (13 classes) 45.9 # 23

Methods


No methods listed for this paper. Add relevant methods here