ConsNet: Learning Consistency Graph for Zero-Shot Human-Object Interaction Detection

14 Aug 2020  ·  Ye Liu, Junsong Yuan, Chang Wen Chen ·

We consider the problem of Human-Object Interaction (HOI) Detection, which aims to locate and recognize HOI instances in the form of <human, action, object> in images. Most existing works treat HOIs as individual interaction categories, thus can not handle the problem of long-tail distribution and polysemy of action labels. We argue that multi-level consistencies among objects, actions and interactions are strong cues for generating semantic representations of rare or previously unseen HOIs. Leveraging the compositional and relational peculiarities of HOI labels, we propose ConsNet, a knowledge-aware framework that explicitly encodes the relations among objects, actions and interactions into an undirected graph called consistency graph, and exploits Graph Attention Networks (GATs) to propagate knowledge among HOI categories as well as their constituents. Our model takes visual features of candidate human-object pairs and word embeddings of HOI labels as inputs, maps them into visual-semantic joint embedding space and obtains detection results by measuring their similarities. We extensively evaluate our model on the challenging V-COCO and HICO-DET datasets, and results validate that our approach outperforms state-of-the-arts under both fully-supervised and zero-shot settings. Code is available at https://github.com/yeliudev/ConsNet.

PDF Abstract

Datasets


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Zero-Shot Human-Object Interaction Detection HICO-DET ConsNet (ResNet-50) mAP (UC) 19.81 # 3
mAP (UO) 20.71 # 2
mAP (UA) 19.04 # 1
Human-Object Interaction Detection HICO-DET ConsNet-F (ResNet-50) mAP 25.94 # 33
Human-Object Interaction Detection HICO-DET ConsNet (ResNet-50) mAP 22.15 # 40

Methods


No methods listed for this paper. Add relevant methods here