Weakly-Supervised Action Localization by Generative Attention Modeling

CVPR 2020  ·  Baifeng Shi, Qi Dai, Yadong Mu, Jingdong Wang ·

Weakly-supervised temporal action localization is a problem of learning an action localization model with only video-level action labeling available. The general framework largely relies on the classification activation, which employs an attention model to identify the action-related frames and then categorizes them into different classes. Such method results in the action-context confusion issue: context frames near action clips tend to be recognized as action frames themselves, since they are closely related to the specific classes. To solve the problem, in this paper we propose to model the class-agnostic frame-wise probability conditioned on the frame attention using conditional Variational Auto-Encoder (VAE). With the observation that the context exhibits notable difference from the action at representation level, a probabilistic model, i.e., conditional VAE, is learned to model the likelihood of each frame given the attention. By maximizing the conditional probability with respect to the attention, the action and non-action frames are well separated. Experiments on THUMOS14 and ActivityNet1.2 demonstrate advantage of our method and effectiveness in handling action-context confusion problem. Code is now available on GitHub.

PDF Abstract CVPR 2020 PDF CVPR 2020 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Weakly Supervised Action Localization ActivityNet-1.2 DGAM mAP@0.5 41.0 # 7
Mean mAP 24.4 # 8
Weakly Supervised Action Localization THUMOS 2014 DGAM mAP@0.5 28.8 # 15
mAP@0.1:0.7 37.0 # 16
mAP@0.1:0.5 45.6 # 15

Methods