Real-time Holistic Robot Pose Estimation with Unknown States

8 Feb 2024  Â·  Shikun Ban, Juling Fan, Wentao Zhu, Xiaoxuan Ma, Yu Qiao, Yizhou Wang ·

Estimating robot pose from RGB images is a crucial problem in computer vision and robotics. While previous methods have achieved promising performance, most of them presume full knowledge of robot internal states, e.g. ground-truth robot joint angles, which are not always available in real-world scenarios. On the other hand, existing approaches that estimate robot pose without joint state priors suffer from heavy computation burdens and thus cannot support real-time applications. This work addresses the urgent need for efficient robot pose estimation with unknown states. We propose an end-to-end pipeline for real-time, holistic robot pose estimation from a single RGB image, even in the absence of known robot states. Our method decomposes the problem into estimating camera-to-robot rotation, robot state parameters, keypoint locations, and root depth. We further design a corresponding neural network module for each task. This approach allows for learning multi-facet representations and facilitates sim-to-real transfer through self-supervised learning. Notably, our method achieves inference with a single feedforward, eliminating the need for costly test-time iterative optimization. As a result, it delivers a 12-time speed boost with state-of-the-art accuracy, enabling real-time holistic robot pose estimation for the first time. Code is available at https://oliverbansk.github.io/Holistic-Robot-Pose/.

PDF Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Robot Pose Estimation DREAM-dataset Holistic-Robot-Pose (known-joint) AUC (avg. on 4 real DREAM datasets) 87.7 # 1
mean-ADD (avg. on 4 real DREAM datasets) 15.0 # 1
Robot Pose Estimation DREAM-dataset Holistic-Robot-Pose (unknown-joint) AUC (avg. on 4 real DREAM datasets) 77.2 # 4
mean-ADD (avg. on 4 real DREAM datasets) 23.1 # 4

Methods


No methods listed for this paper. Add relevant methods here