Asymmetric Cross-Guided Attention Network for Actor and Action Video Segmentation From Natural Language Query

Actor and action video segmentation from natural language query aims to selectively segment the actor and its action in a video based on an input textual description. Previous works mostly focus on learning simple correlation between two heterogeneous features of vision and language via dynamic convolution or fully convolutional classification. However, they ignore the linguistic variation of natural language query and have difficulty in modeling global visual context, which leads to unsatisfactory segmentation performance. To address these issues, we propose an asymmetric cross-guided attention network for actor and action video segmentation from natural language query. Specifically, we frame an asymmetric cross-guided attention network, which consists of vision guided language attention to reduce the linguistic variation of input query and language guided vision attention to incorporate query-focused global visual context simultaneously. Moreover, we adopt multi-resolution fusion scheme and weighted loss for foreground and background pixels to obtain further performance improvement. Extensive experiments on Actor-Action Dataset Sentences and J-HMDB Sentences show that our proposed approach notably outperforms state-of-the-art methods.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Referring Expression Segmentation J-HMDB ACGA Precision@0.5 0.756 # 13
Precision@0.6 0.564 # 16
Precision@0.7 0.287 # 16
Precision@0.8 0.034 # 18
Precision@0.9 0.000 # 11
AP 0.289 # 12
IoU overall 0.576 # 14
IoU mean 0.584 # 11

Results from Other Papers


Task Dataset Model Metric Name Metric Value Rank Source Paper Compare
Referring Expression Segmentation A2D Sentences ACGA Precision@0.5 0.557 # 20
Precision@0.9 0.02 # 23
IoU overall 0.601 # 21
IoU mean 0.490 # 21
Precision@0.6 0.459 # 20
Precision@0.7 0.319 # 22
Precision@0.8 0.16 # 22
AP 0.274 # 18

Methods