Mutual Modality Learning for Video Action Classification

4 Nov 2020  ·  Stepan Komkov, Maksim Dzabraev, Aleksandr Petiushko ·

The construction of models for video action classification progresses rapidly. However, the performance of those models can still be easily improved by ensembling with the same models trained on different modalities (e.g. Optical flow). Unfortunately, it is computationally expensive to use several modalities during inference. Recent works examine the ways to integrate advantages of multi-modality into a single RGB-model. Yet, there is still a room for improvement. In this paper, we explore the various methods to embed the ensemble power into a single model. We show that proper initialization, as well as mutual modality learning, enhances single-modality models. As a result, we achieve state-of-the-art results in the Something-Something-v2 benchmark.

PDF Abstract

Results from the Paper


Ranked #47 on Action Recognition on Something-Something V2 (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Action Recognition Something-Something V2 MML (ensemble) Top-1 Accuracy 69.02 # 47
Top-5 Accuracy 92.70 # 24
Action Recognition Something-Something V2 MML (single) Top-1 Accuracy 66.83 # 73
Top-5 Accuracy 91.30 # 38

Methods


No methods listed for this paper. Add relevant methods here