Multi-Person 3D Motion Prediction with Multi-Range Transformers

We propose a novel framework for multi-person 3D motion trajectory prediction. Our key observation is that a human's action and behaviors may highly depend on the other persons around. Thus, instead of predicting each human pose trajectory in isolation, we introduce a Multi-Range Transformers model which contains of a local-range encoder for individual motion and a global-range encoder for social interactions. The Transformer decoder then performs prediction for each person by taking a corresponding pose as a query which attends to both local and global-range encoder features. Our model not only outperforms state-of-the-art methods on long-term 3D motion prediction, but also generates diverse social interactions. More interestingly, our model can even predict 15-person motion simultaneously by automatically dividing the persons into different interaction groups. Project page with code is available at https://jiashunwang.github.io/MRT/.

PDF Abstract NeurIPS 2021 PDF NeurIPS 2021 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Multi-Person Pose forecasting Expi - common actions split MRT Average MPJPE (mm) @ 1000 ms 238 # 3
Average MPJPE (mm) @ 600 ms 163 # 4
Average MPJPE (mm) @ 400 ms 116 # 4
Average MPJPE (mm) @ 200 ms 58 # 4
Multi-Person Pose forecasting Expi - unseen actions split MRT Average MPJPE (mm) @ 800 ms 291 # 5
Average MPJPE (mm) @ 600 ms 205 # 4
Average MPJPE (mm) @ 400 ms 146 # 4

Methods