The Pose Knows: Video Forecasting by Generating Pose Futures

Current approaches in video forecasting attempt to generate videos directly in pixel space using Generative Adversarial Networks (GANs) or Variational Autoencoders (VAEs). However, since these approaches try to model all the structure and scene dynamics at once, in unconstrained settings they often generate uninterpretable results. Our insight is to model the forecasting problem at a higher level of abstraction. Specifically, we exploit human pose detectors as a free source of supervision and break the video forecasting problem into two discrete steps. First we explicitly model the high level structure of active objects in the scene---humans---and use a VAE to model the possible future movements of humans in the pose space. We then use the future poses generated as conditional information to a GAN to predict the future frames of the video in pixel space. By using the structured space of pose as an intermediate representation, we sidestep the problems that GANs have in generating video pixels directly. We show through quantitative and qualitative evaluation that our method outperforms state-of-the-art methods for video prediction.

PDF Abstract ICCV 2017 PDF ICCV 2017 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Human Pose Forecasting AMASS ThePoseKnows ADE 0.656 # 4
FDE 0.675 # 4
APD 9.283 # 4

Results from Other Papers


Task Dataset Model Metric Name Metric Value Rank Source Paper Compare
Human Pose Forecasting Human3.6M Pose-Knows APD 6723 # 9
ADE 461 # 8
FDE 560 # 8
MMADE 522 # 7
MMFDE 569 # 8
CMD 6.326 # 2
FID 0.538 # 2
Human Pose Forecasting HumanEva-I Pose-Knows APD@2000ms 2308 # 8
ADE@2000ms 269 # 5
FDE@2000ms 296 # 7

Methods