Learning Fast and Robust Target Models for Video Object Segmentation

Video object segmentation (VOS) is a highly challenging problem since the initial mask, defining the target object, is only given at test-time. The main difficulty is to effectively handle appearance changes and similar background objects, while maintaining accurate segmentation. Most previous approaches fine-tune segmentation networks on the first frame, resulting in impractical frame-rates and risk of overfitting. More recent methods integrate generative target appearance models, but either achieve limited robustness or require large amounts of training data. We propose a novel VOS architecture consisting of two network components. The target appearance model consists of a light-weight module, which is learned during the inference stage using fast optimization techniques to predict a coarse but robust target segmentation. The segmentation model is exclusively trained offline, designed to process the coarse scores into high quality segmentation masks. Our method is fast, easily trainable and remains highly effective in cases of limited training data. We perform extensive experiments on the challenging YouTube-VOS and DAVIS datasets. Our network achieves favorable performance, while operating at higher frame-rates compared to state-of-the-art. Code and trained models are available at https://github.com/andr345/frtm-vos.

PDF Abstract CVPR 2020 PDF CVPR 2020 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Semi-Supervised Video Object Segmentation DAVIS 2016 FRTM (val) J&F 81.7 # 58
Speed (FPS) 21.9 # 22
Semi-Supervised Video Object Segmentation DAVIS (no YouTube-VOS training) FRTM FPS 21.9 # 9
D16 val (G) 81.7 # 14
D17 val (G) 68.8 # 20
D17 val (J) 66.4 # 19
D17 val (F) 71.2 # 21
Semi-Supervised Video Object Segmentation YouTube-VOS 2018 FRTM F-Measure (Seen) 76.2 # 43
F-Measure (Unseen) 74.1 # 42
Overall 72.1 # 42
Speed (FPS) 65.9 # 1
Jaccard (Seen) 72.3 # 42

Methods