Asymmetric Loss For Multi-Label Classification

In a typical multi-label setting, a picture contains on average few positive labels, and many negative ones. This positive-negative imbalance dominates the optimization process, and can lead to under-emphasizing gradients from positive labels during training, resulting in poor accuracy. In this paper, we introduce a novel asymmetric loss ("ASL"), which operates differently on positive and negative samples. The loss enables to dynamically down-weights and hard-thresholds easy negative samples, while also discarding possibly mislabeled samples. We demonstrate how ASL can balance the probabilities of different samples, and how this balancing is translated to better mAP scores. With ASL, we reach state-of-the-art results on multiple popular multi-label datasets: MS-COCO, Pascal-VOC, NUS-WIDE and Open Images. We also demonstrate ASL applicability for other tasks, such as single-label classification and object detection. ASL is effective, easy to implement, and does not increase the training time or complexity. Implementation is available at: https://github.com/Alibaba-MIIL/ASL.

PDF Abstract ICCV 2021 PDF ICCV 2021 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Multi-Label Classification MS-COCO TResNet-XL (resolution 640) mAP 88.4 # 15
Multi-Label Classification MS-COCO TResNet-L (resolution 448) mAP 86.6 # 18
Multi-Label Classification NUS-WIDE TResNet-L (resolution 448) MAP 65.2 # 4
Multi-Label Classification OpenImages-v6 TResNet-L mAP 86.3 # 4
Multi-Label Classification PASCAL VOC 2007 TResNet-L (resolution 448, pretrain from MS-COCO) mAP 95.8 # 7
Multi-Label Classification PASCAL VOC 2007 TResNet-L (resolution 448, pretrain from ImageNet) mAP 94.6 # 11

Methods


No methods listed for this paper. Add relevant methods here