Hierarchical Average Precision Training for Pertinent Image Retrieval

5 Jul 2022  ยท  Elias Ramzi, Nicolas Audebert, Nicolas Thome, Clรฉment Rambour, Xavier Bitot ยท

Image Retrieval is commonly evaluated with Average Precision (AP) or Recall@k. Yet, those metrics, are limited to binary labels and do not take into account errors' severity. This paper introduces a new hierarchical AP training method for pertinent image retrieval (HAP-PIER). HAPPIER is based on a new H-AP metric, which leverages a concept hierarchy to refine AP by integrating errors' importance and better evaluate rankings. To train deep models with H-AP, we carefully study the problem's structure and design a smooth lower bound surrogate combined with a clustering loss that ensures consistent ordering. Extensive experiments on 6 datasets show that HAPPIER significantly outperforms state-of-the-art methods for hierarchical retrieval, while being on par with the latest approaches when evaluating fine-grained ranking performances. Finally, we show that HAPPIER leads to better organization of the embedding space, and prevents most severe failure cases of non-hierarchical methods. Our code is publicly available at: https://github.com/elias-ramzi/HAPPIER.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Metric Learning DyML-Animal HAPPIER Average-mAP 43.8 # 1
Metric Learning DyML-Product HAPPIER Average-mAP 38.0 # 1
Metric Learning DyML-Vehicle HAPPIER Average-mAP 37.0 # 1
Image Retrieval iNaturalist HAPPIER (ResNet-50) R@1 70.7 # 7
Image Retrieval iNaturalist HAPPIER_F (ResNet-50) R@1 71.0 # 6
Metric Learning Stanford Online Products HAPPIER_F R@1 81.8 # 13
Metric Learning Stanford Online Products HAPPIER R@1 81.0 # 19

Methods