PhraseCut: Language-based Image Segmentation in the Wild

We consider the problem of segmenting image regions given a natural language phrase, and study it on a novel dataset of 77,262 images and 345,486 phrase-region pairs. Our dataset is collected on top of the Visual Genome dataset and uses the existing annotations to generate a challenging set of referring phrases for which the corresponding regions are manually annotated. Phrases in our dataset correspond to multiple regions and describe a large number of object and stuff categories as well as their attributes such as color, shape, parts, and relationships with other entities in the image. Our experiments show that the scale and diversity of concepts in our dataset poses significant challenges to the existing state-of-the-art. We systematically handle the long-tail nature of these concepts and present a modular approach to combine category, attribute, and relationship cues that outperforms existing approaches.

PDF Abstract CVPR 2020 PDF CVPR 2020 Abstract

Datasets


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Referring Expression Segmentation PhraseCut HULANet Mean IoU 41.3 # 4
Pr@0.5 42.9 # 2
Pr@0.7 27.8 # 2
Pr@0.9 5.9 # 2
Referring Expression Segmentation PhraseCut MattNet Mean IoU 20.2 # 6
Pr@0.5 19.7 # 4
Pr@0.7 13.5 # 3
Pr@0.9 3 # 3
Referring Expression Segmentation PhraseCut RMI Mean IoU 21.1 # 5
Pr@0.5 22 # 3
Pr@0.7 11.6 # 4
Pr@0.9 1.5 # 4

Methods


No methods listed for this paper. Add relevant methods here