AutoFocusFormer: Image Segmentation off the Grid

Real world images often have highly imbalanced content density. Some areas are very uniform, e.g., large patches of blue sky, while other areas are scattered with many small objects. Yet, the commonly used successive grid downsampling strategy in convolutional deep networks treats all areas equally. Hence, small objects are represented in very few spatial locations, leading to worse results in tasks such as segmentation. Intuitively, retaining more pixels representing small objects during downsampling helps to preserve important information. To achieve this, we propose AutoFocusFormer (AFF), a local-attention transformer image recognition backbone, which performs adaptive downsampling by learning to retain the most important pixels for the task. Since adaptive downsampling generates a set of pixels irregularly distributed on the image plane, we abandon the classic grid structure. Instead, we develop a novel point-based local attention block, facilitated by a balanced clustering module and a learnable neighborhood merging module, which yields representations for our point-based versions of state-of-the-art segmentation heads. Experiments show that our AutoFocusFormer (AFF) improves significantly over baseline models of similar sizes.

PDF Abstract CVPR 2023 PDF CVPR 2023 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Instance Segmentation Cityscapes val AFF-Base (single-scale, point-based Mask2Former) mask AP 46.2 # 4
AP50 74.2 # 1
Panoptic Segmentation Cityscapes val AFF-Base (single-scale, point-based Mask2Former) PQ 67.7 # 8
PQst 71.5 # 2
PQth 62.5 # 3
mIoU 83.0 # 10
AP 46.2 # 5
Instance Segmentation Cityscapes val AFF-Small (single-scale, point-based Mask2Former) mask AP 44.0 # 8
AP50 72.8 # 2
Panoptic Segmentation Cityscapes val AFF-Small (single-scale, point-based Mask2Former) PQ 66.9 # 13
PQst 70.8 # 3
PQth 61.5 # 4
mIoU 82.2 # 14
AP 44.2 # 9

Methods


No methods listed for this paper. Add relevant methods here