Towards Universal Representation for Unseen Action Recognition

CVPR 2018  ·  Yi Zhu, Yang Long, Yu Guan, Shawn Newsam, Ling Shao ·

Unseen Action Recognition (UAR) aims to recognise novel action categories without training examples. While previous methods focus on inner-dataset seen/unseen splits, this paper proposes a pipeline using a large-scale training source to achieve a Universal Representation (UR) that can generalise to a more realistic Cross-Dataset UAR (CD-UAR) scenario. We first address UAR as a Generalised Multiple-Instance Learning (GMIL) problem and discover 'building-blocks' from the large-scale ActivityNet dataset using distribution kernels. Essential visual and semantic components are preserved in a shared space to achieve the UR that can efficiently generalise to new datasets. Predicted UR exemplars can be improved by a simple semantic adaptation, and then an unseen action can be directly recognised using UR during the test. Without further training, extensive experiments manifest significant improvements over the UCF101 and HMDB51 benchmarks.

PDF Abstract CVPR 2018 PDF CVPR 2018 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Action Recognition ActivityNet CD-UAR mAP 53.8 # 14
Zero-Shot Action Recognition HMDB51 UR Top-1 Accuracy 24.4 # 19
Action Recognition HMDB-51 CD-UAR Average accuracy of 3 splits 51.8 # 74
Zero-Shot Action Recognition UCF101 UR Top-1 Accuracy 17.5 # 23
Action Recognition UCF101 CD-UAR 3-fold Accuracy 42.5 # 85

Methods


No methods listed for this paper. Add relevant methods here