MOVE: Unsupervised Movable Object Segmentation and Detection

14 Oct 2022  ·  Adam Bielski, Paolo Favaro ·

We introduce MOVE, a novel method to segment objects without any form of supervision. MOVE exploits the fact that foreground objects can be shifted locally relative to their initial position and result in realistic (undistorted) new images. This property allows us to train a segmentation model on a dataset of images without annotation and to achieve state of the art (SotA) performance on several evaluation datasets for unsupervised salient object detection and segmentation. In unsupervised single object discovery, MOVE gives an average CorLoc improvement of 7.2% over the SotA, and in unsupervised class-agnostic object detection it gives a relative AP improvement of 53% on average. Our approach is built on top of self-supervised features (e.g. from DINO or MAE), an inpainting network (based on the Masked AutoEncoder) and adversarial training.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Single-object discovery COCO_20k MOVE CorLoc 66.6 # 3
Single-object discovery COCO_20k MOVE + CAD CorLoc 71.9 # 2
Unsupervised Saliency Detection DUT-OMRON MOVE Accuracy 93.7 # 1
IoU 66.6 # 1
maximal F-measure 76.6 # 2
Unsupervised Saliency Detection DUTS MOVE Accuracy 95.4 # 1
IoU 72.8 # 1
maximal F-measure 82.9 # 2
Unsupervised Saliency Detection ECSSD MOVE Accuracy 95.6 # 1
IoU 83.6 # 1
maximal F-measure 92.1 # 2

Methods